00:00:00.001 Started by upstream project "autotest-per-patch" build number 132083 00:00:00.001 originally caused by: 00:00:00.001 Started by user sys_sgci 00:00:00.059 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-phy.groovy 00:00:00.060 The recommended git tool is: git 00:00:00.060 using credential 00000000-0000-0000-0000-000000000002 00:00:00.062 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.109 Fetching changes from the remote Git repository 00:00:00.111 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.164 Using shallow fetch with depth 1 00:00:00.164 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.164 > git --version # timeout=10 00:00:00.216 > git --version # 'git version 2.39.2' 00:00:00.216 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.245 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.246 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:05.210 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:05.222 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:05.235 Checking out Revision b9dd3f7ec12b0ee8a44940dc99ce739345caa4cf (FETCH_HEAD) 00:00:05.235 > git config core.sparsecheckout # timeout=10 00:00:05.245 > git read-tree -mu HEAD # timeout=10 00:00:05.261 > git checkout -f b9dd3f7ec12b0ee8a44940dc99ce739345caa4cf # timeout=5 00:00:05.282 Commit message: "jenkins/jjb-config: Ignore OS version mismatch under freebsd" 00:00:05.282 > git rev-list --no-walk b9dd3f7ec12b0ee8a44940dc99ce739345caa4cf # timeout=10 00:00:05.464 [Pipeline] Start of Pipeline 00:00:05.478 [Pipeline] library 00:00:05.480 Loading library shm_lib@master 00:00:05.480 Library shm_lib@master is cached. Copying from home. 00:00:05.496 [Pipeline] node 00:00:05.513 Running on CYP9 in /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:05.515 [Pipeline] { 00:00:05.526 [Pipeline] catchError 00:00:05.528 [Pipeline] { 00:00:05.541 [Pipeline] wrap 00:00:05.549 [Pipeline] { 00:00:05.557 [Pipeline] stage 00:00:05.559 [Pipeline] { (Prologue) 00:00:05.805 [Pipeline] sh 00:00:06.093 + logger -p user.info -t JENKINS-CI 00:00:06.114 [Pipeline] echo 00:00:06.116 Node: CYP9 00:00:06.123 [Pipeline] sh 00:00:06.433 [Pipeline] setCustomBuildProperty 00:00:06.446 [Pipeline] echo 00:00:06.448 Cleanup processes 00:00:06.453 [Pipeline] sh 00:00:06.743 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:06.743 1564529 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:06.758 [Pipeline] sh 00:00:07.047 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:07.047 ++ grep -v 'sudo pgrep' 00:00:07.047 ++ awk '{print $1}' 00:00:07.047 + sudo kill -9 00:00:07.047 + true 00:00:07.064 [Pipeline] cleanWs 00:00:07.075 [WS-CLEANUP] Deleting project workspace... 00:00:07.075 [WS-CLEANUP] Deferred wipeout is used... 00:00:07.083 [WS-CLEANUP] done 00:00:07.087 [Pipeline] setCustomBuildProperty 00:00:07.102 [Pipeline] sh 00:00:07.389 + sudo git config --global --replace-all safe.directory '*' 00:00:07.492 [Pipeline] httpRequest 00:00:07.860 [Pipeline] echo 00:00:07.862 Sorcerer 10.211.164.101 is alive 00:00:07.873 [Pipeline] retry 00:00:07.875 [Pipeline] { 00:00:07.889 [Pipeline] httpRequest 00:00:07.894 HttpMethod: GET 00:00:07.894 URL: http://10.211.164.101/packages/jbp_b9dd3f7ec12b0ee8a44940dc99ce739345caa4cf.tar.gz 00:00:07.895 Sending request to url: http://10.211.164.101/packages/jbp_b9dd3f7ec12b0ee8a44940dc99ce739345caa4cf.tar.gz 00:00:07.923 Response Code: HTTP/1.1 200 OK 00:00:07.924 Success: Status code 200 is in the accepted range: 200,404 00:00:07.924 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/jbp_b9dd3f7ec12b0ee8a44940dc99ce739345caa4cf.tar.gz 00:00:22.818 [Pipeline] } 00:00:22.835 [Pipeline] // retry 00:00:22.843 [Pipeline] sh 00:00:23.128 + tar --no-same-owner -xf jbp_b9dd3f7ec12b0ee8a44940dc99ce739345caa4cf.tar.gz 00:00:23.142 [Pipeline] httpRequest 00:00:24.505 [Pipeline] echo 00:00:24.506 Sorcerer 10.211.164.101 is alive 00:00:24.511 [Pipeline] retry 00:00:24.514 [Pipeline] { 00:00:24.524 [Pipeline] httpRequest 00:00:24.528 HttpMethod: GET 00:00:24.528 URL: http://10.211.164.101/packages/spdk_15b283ee8346c97ff179d6c052026f7957f85e30.tar.gz 00:00:24.529 Sending request to url: http://10.211.164.101/packages/spdk_15b283ee8346c97ff179d6c052026f7957f85e30.tar.gz 00:00:24.536 Response Code: HTTP/1.1 200 OK 00:00:24.536 Success: Status code 200 is in the accepted range: 200,404 00:00:24.537 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk_15b283ee8346c97ff179d6c052026f7957f85e30.tar.gz 00:04:59.753 [Pipeline] } 00:04:59.764 [Pipeline] // retry 00:04:59.770 [Pipeline] sh 00:05:00.054 + tar --no-same-owner -xf spdk_15b283ee8346c97ff179d6c052026f7957f85e30.tar.gz 00:05:03.364 [Pipeline] sh 00:05:03.649 + git -C spdk log --oneline -n5 00:05:03.650 15b283ee8 nvmf: Get metadata config by not bdev but bdev_desc 00:05:03.650 cec609db6 bdevperf: Add no_metadata option 00:05:03.650 39e719aa5 bdevperf: Get metadata config by not bdev but bdev_desc 00:05:03.650 a32bce303 bdevperf: g_main_thread calls bdev_open() instead of job->thread 00:05:03.650 724236955 bdev/malloc: Fix unexpected DIF verification error for initial read 00:05:03.660 [Pipeline] } 00:05:03.673 [Pipeline] // stage 00:05:03.681 [Pipeline] stage 00:05:03.684 [Pipeline] { (Prepare) 00:05:03.699 [Pipeline] writeFile 00:05:03.712 [Pipeline] sh 00:05:03.998 + logger -p user.info -t JENKINS-CI 00:05:04.010 [Pipeline] sh 00:05:04.295 + logger -p user.info -t JENKINS-CI 00:05:04.306 [Pipeline] sh 00:05:04.591 + cat autorun-spdk.conf 00:05:04.591 SPDK_RUN_FUNCTIONAL_TEST=1 00:05:04.591 SPDK_TEST_NVMF=1 00:05:04.591 SPDK_TEST_NVME_CLI=1 00:05:04.591 SPDK_TEST_NVMF_TRANSPORT=tcp 00:05:04.591 SPDK_TEST_NVMF_NICS=e810 00:05:04.591 SPDK_TEST_VFIOUSER=1 00:05:04.591 SPDK_RUN_UBSAN=1 00:05:04.591 NET_TYPE=phy 00:05:04.599 RUN_NIGHTLY=0 00:05:04.604 [Pipeline] readFile 00:05:04.628 [Pipeline] withEnv 00:05:04.630 [Pipeline] { 00:05:04.643 [Pipeline] sh 00:05:04.931 + set -ex 00:05:04.931 + [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf ]] 00:05:04.931 + source /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:05:04.931 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:05:04.931 ++ SPDK_TEST_NVMF=1 00:05:04.931 ++ SPDK_TEST_NVME_CLI=1 00:05:04.931 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:05:04.931 ++ SPDK_TEST_NVMF_NICS=e810 00:05:04.931 ++ SPDK_TEST_VFIOUSER=1 00:05:04.931 ++ SPDK_RUN_UBSAN=1 00:05:04.931 ++ NET_TYPE=phy 00:05:04.931 ++ RUN_NIGHTLY=0 00:05:04.931 + case $SPDK_TEST_NVMF_NICS in 00:05:04.931 + DRIVERS=ice 00:05:04.931 + [[ tcp == \r\d\m\a ]] 00:05:04.931 + [[ -n ice ]] 00:05:04.931 + sudo rmmod mlx4_ib mlx5_ib irdma i40iw iw_cxgb4 00:05:04.931 rmmod: ERROR: Module mlx4_ib is not currently loaded 00:05:04.931 rmmod: ERROR: Module mlx5_ib is not currently loaded 00:05:04.931 rmmod: ERROR: Module irdma is not currently loaded 00:05:04.931 rmmod: ERROR: Module i40iw is not currently loaded 00:05:04.931 rmmod: ERROR: Module iw_cxgb4 is not currently loaded 00:05:04.931 + true 00:05:04.931 + for D in $DRIVERS 00:05:04.931 + sudo modprobe ice 00:05:04.931 + exit 0 00:05:04.941 [Pipeline] } 00:05:04.956 [Pipeline] // withEnv 00:05:04.961 [Pipeline] } 00:05:04.973 [Pipeline] // stage 00:05:04.981 [Pipeline] catchError 00:05:04.983 [Pipeline] { 00:05:04.996 [Pipeline] timeout 00:05:04.996 Timeout set to expire in 1 hr 0 min 00:05:04.998 [Pipeline] { 00:05:05.011 [Pipeline] stage 00:05:05.013 [Pipeline] { (Tests) 00:05:05.027 [Pipeline] sh 00:05:05.315 + jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:05:05.315 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:05:05.315 + DIR_ROOT=/var/jenkins/workspace/nvmf-tcp-phy-autotest 00:05:05.315 + [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest ]] 00:05:05.315 + DIR_SPDK=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:05:05.315 + DIR_OUTPUT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:05:05.315 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk ]] 00:05:05.315 + [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:05:05.315 + mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:05:05.315 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:05:05.315 + [[ nvmf-tcp-phy-autotest == pkgdep-* ]] 00:05:05.315 + cd /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:05:05.315 + source /etc/os-release 00:05:05.315 ++ NAME='Fedora Linux' 00:05:05.315 ++ VERSION='39 (Cloud Edition)' 00:05:05.315 ++ ID=fedora 00:05:05.315 ++ VERSION_ID=39 00:05:05.315 ++ VERSION_CODENAME= 00:05:05.315 ++ PLATFORM_ID=platform:f39 00:05:05.315 ++ PRETTY_NAME='Fedora Linux 39 (Cloud Edition)' 00:05:05.315 ++ ANSI_COLOR='0;38;2;60;110;180' 00:05:05.315 ++ LOGO=fedora-logo-icon 00:05:05.315 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:39 00:05:05.315 ++ HOME_URL=https://fedoraproject.org/ 00:05:05.315 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f39/system-administrators-guide/ 00:05:05.315 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:05:05.315 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:05:05.315 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:05:05.315 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=39 00:05:05.315 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:05:05.315 ++ REDHAT_SUPPORT_PRODUCT_VERSION=39 00:05:05.315 ++ SUPPORT_END=2024-11-12 00:05:05.315 ++ VARIANT='Cloud Edition' 00:05:05.315 ++ VARIANT_ID=cloud 00:05:05.315 + uname -a 00:05:05.315 Linux spdk-cyp-09 6.8.9-200.fc39.x86_64 #1 SMP PREEMPT_DYNAMIC Wed Jul 24 03:04:40 UTC 2024 x86_64 GNU/Linux 00:05:05.315 + sudo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:05:08.616 Hugepages 00:05:08.616 node hugesize free / total 00:05:08.616 node0 1048576kB 0 / 0 00:05:08.616 node0 2048kB 0 / 0 00:05:08.616 node1 1048576kB 0 / 0 00:05:08.616 node1 2048kB 0 / 0 00:05:08.616 00:05:08.616 Type BDF Vendor Device NUMA Driver Device Block devices 00:05:08.616 I/OAT 0000:00:01.0 8086 0b00 0 ioatdma - - 00:05:08.616 I/OAT 0000:00:01.1 8086 0b00 0 ioatdma - - 00:05:08.616 I/OAT 0000:00:01.2 8086 0b00 0 ioatdma - - 00:05:08.616 I/OAT 0000:00:01.3 8086 0b00 0 ioatdma - - 00:05:08.616 I/OAT 0000:00:01.4 8086 0b00 0 ioatdma - - 00:05:08.616 I/OAT 0000:00:01.5 8086 0b00 0 ioatdma - - 00:05:08.616 I/OAT 0000:00:01.6 8086 0b00 0 ioatdma - - 00:05:08.617 I/OAT 0000:00:01.7 8086 0b00 0 ioatdma - - 00:05:08.617 NVMe 0000:65:00.0 144d a80a 0 nvme nvme0 nvme0n1 00:05:08.617 I/OAT 0000:80:01.0 8086 0b00 1 ioatdma - - 00:05:08.617 I/OAT 0000:80:01.1 8086 0b00 1 ioatdma - - 00:05:08.617 I/OAT 0000:80:01.2 8086 0b00 1 ioatdma - - 00:05:08.617 I/OAT 0000:80:01.3 8086 0b00 1 ioatdma - - 00:05:08.617 I/OAT 0000:80:01.4 8086 0b00 1 ioatdma - - 00:05:08.617 I/OAT 0000:80:01.5 8086 0b00 1 ioatdma - - 00:05:08.617 I/OAT 0000:80:01.6 8086 0b00 1 ioatdma - - 00:05:08.617 I/OAT 0000:80:01.7 8086 0b00 1 ioatdma - - 00:05:08.617 + rm -f /tmp/spdk-ld-path 00:05:08.617 + source autorun-spdk.conf 00:05:08.617 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:05:08.617 ++ SPDK_TEST_NVMF=1 00:05:08.617 ++ SPDK_TEST_NVME_CLI=1 00:05:08.617 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:05:08.617 ++ SPDK_TEST_NVMF_NICS=e810 00:05:08.617 ++ SPDK_TEST_VFIOUSER=1 00:05:08.617 ++ SPDK_RUN_UBSAN=1 00:05:08.617 ++ NET_TYPE=phy 00:05:08.617 ++ RUN_NIGHTLY=0 00:05:08.617 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:05:08.617 + [[ -n '' ]] 00:05:08.617 + sudo git config --global --add safe.directory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:05:08.617 + for M in /var/spdk/build-*-manifest.txt 00:05:08.617 + [[ -f /var/spdk/build-kernel-manifest.txt ]] 00:05:08.617 + cp /var/spdk/build-kernel-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:05:08.617 + for M in /var/spdk/build-*-manifest.txt 00:05:08.617 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:05:08.617 + cp /var/spdk/build-pkg-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:05:08.617 + for M in /var/spdk/build-*-manifest.txt 00:05:08.617 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:05:08.617 + cp /var/spdk/build-repo-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:05:08.617 ++ uname 00:05:08.617 + [[ Linux == \L\i\n\u\x ]] 00:05:08.617 + sudo dmesg -T 00:05:08.617 + sudo dmesg --clear 00:05:08.617 + dmesg_pid=1566672 00:05:08.617 + [[ Fedora Linux == FreeBSD ]] 00:05:08.617 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:05:08.617 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:05:08.617 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:05:08.617 + [[ -x /usr/src/fio-static/fio ]] 00:05:08.617 + export FIO_BIN=/usr/src/fio-static/fio 00:05:08.617 + FIO_BIN=/usr/src/fio-static/fio 00:05:08.617 + sudo dmesg -Tw 00:05:08.617 + [[ '' == \/\v\a\r\/\j\e\n\k\i\n\s\/\w\o\r\k\s\p\a\c\e\/\n\v\m\f\-\t\c\p\-\p\h\y\-\a\u\t\o\t\e\s\t\/\q\e\m\u\_\v\f\i\o\/* ]] 00:05:08.617 + [[ ! -v VFIO_QEMU_BIN ]] 00:05:08.617 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:05:08.617 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:05:08.617 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:05:08.617 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:05:08.617 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:05:08.617 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:05:08.617 + spdk/autorun.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:05:08.617 Test configuration: 00:05:08.617 SPDK_RUN_FUNCTIONAL_TEST=1 00:05:08.617 SPDK_TEST_NVMF=1 00:05:08.617 SPDK_TEST_NVME_CLI=1 00:05:08.617 SPDK_TEST_NVMF_TRANSPORT=tcp 00:05:08.617 SPDK_TEST_NVMF_NICS=e810 00:05:08.617 SPDK_TEST_VFIOUSER=1 00:05:08.617 SPDK_RUN_UBSAN=1 00:05:08.617 NET_TYPE=phy 00:05:08.617 RUN_NIGHTLY=0 08:40:58 -- common/autotest_common.sh@1688 -- $ [[ n == y ]] 00:05:08.617 08:40:58 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:05:08.617 08:40:58 -- scripts/common.sh@15 -- $ shopt -s extglob 00:05:08.617 08:40:58 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:05:08.617 08:40:58 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:05:08.617 08:40:58 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:05:08.617 08:40:58 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:08.617 08:40:58 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:08.617 08:40:58 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:08.617 08:40:58 -- paths/export.sh@5 -- $ export PATH 00:05:08.617 08:40:58 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:05:08.617 08:40:58 -- common/autobuild_common.sh@485 -- $ out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:05:08.617 08:40:58 -- common/autobuild_common.sh@486 -- $ date +%s 00:05:08.617 08:40:58 -- common/autobuild_common.sh@486 -- $ mktemp -dt spdk_1730878858.XXXXXX 00:05:08.617 08:40:58 -- common/autobuild_common.sh@486 -- $ SPDK_WORKSPACE=/tmp/spdk_1730878858.tu3olK 00:05:08.617 08:40:58 -- common/autobuild_common.sh@488 -- $ [[ -n '' ]] 00:05:08.617 08:40:58 -- common/autobuild_common.sh@492 -- $ '[' -n '' ']' 00:05:08.617 08:40:58 -- common/autobuild_common.sh@495 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/' 00:05:08.617 08:40:58 -- common/autobuild_common.sh@499 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp' 00:05:08.617 08:40:58 -- common/autobuild_common.sh@501 -- $ scanbuild='scan-build -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:05:08.617 08:40:58 -- common/autobuild_common.sh@502 -- $ get_config_params 00:05:08.617 08:40:58 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:05:08.617 08:40:58 -- common/autotest_common.sh@10 -- $ set +x 00:05:08.617 08:40:58 -- common/autobuild_common.sh@502 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user' 00:05:08.617 08:40:58 -- common/autobuild_common.sh@504 -- $ start_monitor_resources 00:05:08.617 08:40:58 -- pm/common@17 -- $ local monitor 00:05:08.617 08:40:58 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:08.617 08:40:58 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:08.617 08:40:58 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:08.617 08:40:58 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:05:08.617 08:40:58 -- pm/common@21 -- $ date +%s 00:05:08.617 08:40:58 -- pm/common@25 -- $ sleep 1 00:05:08.617 08:40:58 -- pm/common@21 -- $ date +%s 00:05:08.617 08:40:58 -- pm/common@21 -- $ date +%s 00:05:08.617 08:40:58 -- pm/common@21 -- $ date +%s 00:05:08.617 08:40:58 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1730878858 00:05:08.617 08:40:58 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1730878858 00:05:08.617 08:40:58 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1730878858 00:05:08.617 08:40:58 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1730878858 00:05:08.617 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1730878858_collect-vmstat.pm.log 00:05:08.617 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1730878858_collect-cpu-load.pm.log 00:05:08.878 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1730878858_collect-cpu-temp.pm.log 00:05:08.878 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1730878858_collect-bmc-pm.bmc.pm.log 00:05:09.818 08:40:59 -- common/autobuild_common.sh@505 -- $ trap stop_monitor_resources EXIT 00:05:09.818 08:40:59 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:05:09.818 08:40:59 -- spdk/autobuild.sh@12 -- $ umask 022 00:05:09.818 08:40:59 -- spdk/autobuild.sh@13 -- $ cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:05:09.818 08:40:59 -- spdk/autobuild.sh@16 -- $ date -u 00:05:09.818 Wed Nov 6 07:40:59 AM UTC 2024 00:05:09.818 08:40:59 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:05:09.818 v25.01-pre-140-g15b283ee8 00:05:09.818 08:40:59 -- spdk/autobuild.sh@19 -- $ '[' 0 -eq 1 ']' 00:05:09.818 08:40:59 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:05:09.818 08:40:59 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:05:09.818 08:40:59 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:05:09.818 08:40:59 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:05:09.818 08:40:59 -- common/autotest_common.sh@10 -- $ set +x 00:05:09.818 ************************************ 00:05:09.818 START TEST ubsan 00:05:09.818 ************************************ 00:05:09.818 08:40:59 ubsan -- common/autotest_common.sh@1125 -- $ echo 'using ubsan' 00:05:09.818 using ubsan 00:05:09.818 00:05:09.818 real 0m0.001s 00:05:09.818 user 0m0.001s 00:05:09.818 sys 0m0.000s 00:05:09.818 08:40:59 ubsan -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:05:09.818 08:40:59 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:05:09.818 ************************************ 00:05:09.818 END TEST ubsan 00:05:09.818 ************************************ 00:05:09.818 08:40:59 -- spdk/autobuild.sh@27 -- $ '[' -n '' ']' 00:05:09.818 08:40:59 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:05:09.818 08:40:59 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:05:09.818 08:40:59 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:05:09.818 08:40:59 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:05:09.818 08:40:59 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:05:09.818 08:40:59 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:05:09.818 08:40:59 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:05:09.818 08:40:59 -- spdk/autobuild.sh@67 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user --with-shared 00:05:10.079 Using default SPDK env in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:05:10.079 Using default DPDK in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:05:10.340 Using 'verbs' RDMA provider 00:05:26.197 Configuring ISA-L (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal.log)...done. 00:05:38.431 Configuring ISA-L-crypto (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal-crypto.log)...done. 00:05:38.431 Creating mk/config.mk...done. 00:05:38.431 Creating mk/cc.flags.mk...done. 00:05:38.431 Type 'make' to build. 00:05:38.431 08:41:28 -- spdk/autobuild.sh@70 -- $ run_test make make -j144 00:05:38.431 08:41:28 -- common/autotest_common.sh@1101 -- $ '[' 3 -le 1 ']' 00:05:38.431 08:41:28 -- common/autotest_common.sh@1107 -- $ xtrace_disable 00:05:38.431 08:41:28 -- common/autotest_common.sh@10 -- $ set +x 00:05:38.431 ************************************ 00:05:38.431 START TEST make 00:05:38.431 ************************************ 00:05:38.431 08:41:28 make -- common/autotest_common.sh@1125 -- $ make -j144 00:05:38.693 make[1]: Nothing to be done for 'all'. 00:05:40.080 The Meson build system 00:05:40.080 Version: 1.5.0 00:05:40.080 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user 00:05:40.080 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:05:40.080 Build type: native build 00:05:40.080 Project name: libvfio-user 00:05:40.080 Project version: 0.0.1 00:05:40.080 C compiler for the host machine: cc (gcc 13.3.1 "cc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:05:40.080 C linker for the host machine: cc ld.bfd 2.40-14 00:05:40.080 Host machine cpu family: x86_64 00:05:40.080 Host machine cpu: x86_64 00:05:40.080 Run-time dependency threads found: YES 00:05:40.080 Library dl found: YES 00:05:40.080 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:05:40.080 Run-time dependency json-c found: YES 0.17 00:05:40.080 Run-time dependency cmocka found: YES 1.1.7 00:05:40.080 Program pytest-3 found: NO 00:05:40.081 Program flake8 found: NO 00:05:40.081 Program misspell-fixer found: NO 00:05:40.081 Program restructuredtext-lint found: NO 00:05:40.081 Program valgrind found: YES (/usr/bin/valgrind) 00:05:40.081 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:05:40.081 Compiler for C supports arguments -Wmissing-declarations: YES 00:05:40.081 Compiler for C supports arguments -Wwrite-strings: YES 00:05:40.081 ../libvfio-user/test/meson.build:20: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:05:40.081 Program test-lspci.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-lspci.sh) 00:05:40.081 Program test-linkage.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-linkage.sh) 00:05:40.081 ../libvfio-user/test/py/meson.build:16: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:05:40.081 Build targets in project: 8 00:05:40.081 WARNING: Project specifies a minimum meson_version '>= 0.53.0' but uses features which were added in newer versions: 00:05:40.081 * 0.57.0: {'exclude_suites arg in add_test_setup'} 00:05:40.081 00:05:40.081 libvfio-user 0.0.1 00:05:40.081 00:05:40.081 User defined options 00:05:40.081 buildtype : debug 00:05:40.081 default_library: shared 00:05:40.081 libdir : /usr/local/lib 00:05:40.081 00:05:40.081 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:05:40.386 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:05:40.386 [1/37] Compiling C object lib/libvfio-user.so.0.0.1.p/irq.c.o 00:05:40.386 [2/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci.c.o 00:05:40.386 [3/37] Compiling C object samples/null.p/null.c.o 00:05:40.386 [4/37] Compiling C object samples/shadow_ioeventfd_server.p/shadow_ioeventfd_server.c.o 00:05:40.386 [5/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran.c.o 00:05:40.386 [6/37] Compiling C object samples/gpio-pci-idio-16.p/gpio-pci-idio-16.c.o 00:05:40.386 [7/37] Compiling C object samples/client.p/.._lib_migration.c.o 00:05:40.386 [8/37] Compiling C object lib/libvfio-user.so.0.0.1.p/migration.c.o 00:05:40.386 [9/37] Compiling C object test/unit_tests.p/.._lib_tran.c.o 00:05:40.386 [10/37] Compiling C object samples/lspci.p/lspci.c.o 00:05:40.386 [11/37] Compiling C object test/unit_tests.p/.._lib_irq.c.o 00:05:40.386 [12/37] Compiling C object test/unit_tests.p/.._lib_migration.c.o 00:05:40.386 [13/37] Compiling C object test/unit_tests.p/.._lib_tran_pipe.c.o 00:05:40.386 [14/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran_sock.c.o 00:05:40.386 [15/37] Compiling C object test/unit_tests.p/mocks.c.o 00:05:40.386 [16/37] Compiling C object test/unit_tests.p/.._lib_pci.c.o 00:05:40.386 [17/37] Compiling C object samples/client.p/.._lib_tran_sock.c.o 00:05:40.386 [18/37] Compiling C object samples/client.p/.._lib_tran.c.o 00:05:40.386 [19/37] Compiling C object lib/libvfio-user.so.0.0.1.p/dma.c.o 00:05:40.386 [20/37] Compiling C object test/unit_tests.p/.._lib_dma.c.o 00:05:40.386 [21/37] Compiling C object test/unit_tests.p/.._lib_pci_caps.c.o 00:05:40.386 [22/37] Compiling C object test/unit_tests.p/.._lib_tran_sock.c.o 00:05:40.387 [23/37] Compiling C object test/unit_tests.p/unit-tests.c.o 00:05:40.387 [24/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci_caps.c.o 00:05:40.387 [25/37] Compiling C object samples/client.p/client.c.o 00:05:40.387 [26/37] Compiling C object samples/server.p/server.c.o 00:05:40.690 [27/37] Linking target samples/client 00:05:40.690 [28/37] Compiling C object test/unit_tests.p/.._lib_libvfio-user.c.o 00:05:40.690 [29/37] Compiling C object lib/libvfio-user.so.0.0.1.p/libvfio-user.c.o 00:05:40.690 [30/37] Linking target lib/libvfio-user.so.0.0.1 00:05:40.690 [31/37] Linking target test/unit_tests 00:05:40.690 [32/37] Generating symbol file lib/libvfio-user.so.0.0.1.p/libvfio-user.so.0.0.1.symbols 00:05:40.690 [33/37] Linking target samples/null 00:05:40.690 [34/37] Linking target samples/server 00:05:40.690 [35/37] Linking target samples/lspci 00:05:40.690 [36/37] Linking target samples/gpio-pci-idio-16 00:05:40.690 [37/37] Linking target samples/shadow_ioeventfd_server 00:05:40.690 INFO: autodetecting backend as ninja 00:05:40.690 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:05:40.962 DESTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user meson install --quiet -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:05:41.224 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:05:41.224 ninja: no work to do. 00:05:47.832 The Meson build system 00:05:47.832 Version: 1.5.0 00:05:47.832 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk 00:05:47.832 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp 00:05:47.832 Build type: native build 00:05:47.832 Program cat found: YES (/usr/bin/cat) 00:05:47.832 Project name: DPDK 00:05:47.832 Project version: 24.03.0 00:05:47.832 C compiler for the host machine: cc (gcc 13.3.1 "cc (GCC) 13.3.1 20240522 (Red Hat 13.3.1-1)") 00:05:47.832 C linker for the host machine: cc ld.bfd 2.40-14 00:05:47.832 Host machine cpu family: x86_64 00:05:47.832 Host machine cpu: x86_64 00:05:47.832 Message: ## Building in Developer Mode ## 00:05:47.832 Program pkg-config found: YES (/usr/bin/pkg-config) 00:05:47.832 Program check-symbols.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/buildtools/check-symbols.sh) 00:05:47.832 Program options-ibverbs-static.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/buildtools/options-ibverbs-static.sh) 00:05:47.832 Program python3 found: YES (/usr/bin/python3) 00:05:47.832 Program cat found: YES (/usr/bin/cat) 00:05:47.832 Compiler for C supports arguments -march=native: YES 00:05:47.832 Checking for size of "void *" : 8 00:05:47.832 Checking for size of "void *" : 8 (cached) 00:05:47.832 Compiler for C supports link arguments -Wl,--undefined-version: YES 00:05:47.832 Library m found: YES 00:05:47.832 Library numa found: YES 00:05:47.832 Has header "numaif.h" : YES 00:05:47.832 Library fdt found: NO 00:05:47.832 Library execinfo found: NO 00:05:47.832 Has header "execinfo.h" : YES 00:05:47.832 Found pkg-config: YES (/usr/bin/pkg-config) 1.9.5 00:05:47.832 Run-time dependency libarchive found: NO (tried pkgconfig) 00:05:47.832 Run-time dependency libbsd found: NO (tried pkgconfig) 00:05:47.832 Run-time dependency jansson found: NO (tried pkgconfig) 00:05:47.832 Run-time dependency openssl found: YES 3.1.1 00:05:47.832 Run-time dependency libpcap found: YES 1.10.4 00:05:47.832 Has header "pcap.h" with dependency libpcap: YES 00:05:47.832 Compiler for C supports arguments -Wcast-qual: YES 00:05:47.832 Compiler for C supports arguments -Wdeprecated: YES 00:05:47.832 Compiler for C supports arguments -Wformat: YES 00:05:47.832 Compiler for C supports arguments -Wformat-nonliteral: NO 00:05:47.832 Compiler for C supports arguments -Wformat-security: NO 00:05:47.832 Compiler for C supports arguments -Wmissing-declarations: YES 00:05:47.832 Compiler for C supports arguments -Wmissing-prototypes: YES 00:05:47.832 Compiler for C supports arguments -Wnested-externs: YES 00:05:47.832 Compiler for C supports arguments -Wold-style-definition: YES 00:05:47.832 Compiler for C supports arguments -Wpointer-arith: YES 00:05:47.832 Compiler for C supports arguments -Wsign-compare: YES 00:05:47.832 Compiler for C supports arguments -Wstrict-prototypes: YES 00:05:47.832 Compiler for C supports arguments -Wundef: YES 00:05:47.832 Compiler for C supports arguments -Wwrite-strings: YES 00:05:47.832 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:05:47.832 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:05:47.832 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:05:47.832 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:05:47.832 Program objdump found: YES (/usr/bin/objdump) 00:05:47.832 Compiler for C supports arguments -mavx512f: YES 00:05:47.832 Checking if "AVX512 checking" compiles: YES 00:05:47.832 Fetching value of define "__SSE4_2__" : 1 00:05:47.832 Fetching value of define "__AES__" : 1 00:05:47.832 Fetching value of define "__AVX__" : 1 00:05:47.832 Fetching value of define "__AVX2__" : 1 00:05:47.832 Fetching value of define "__AVX512BW__" : 1 00:05:47.832 Fetching value of define "__AVX512CD__" : 1 00:05:47.832 Fetching value of define "__AVX512DQ__" : 1 00:05:47.832 Fetching value of define "__AVX512F__" : 1 00:05:47.832 Fetching value of define "__AVX512VL__" : 1 00:05:47.832 Fetching value of define "__PCLMUL__" : 1 00:05:47.832 Fetching value of define "__RDRND__" : 1 00:05:47.832 Fetching value of define "__RDSEED__" : 1 00:05:47.832 Fetching value of define "__VPCLMULQDQ__" : 1 00:05:47.832 Fetching value of define "__znver1__" : (undefined) 00:05:47.832 Fetching value of define "__znver2__" : (undefined) 00:05:47.832 Fetching value of define "__znver3__" : (undefined) 00:05:47.832 Fetching value of define "__znver4__" : (undefined) 00:05:47.832 Compiler for C supports arguments -Wno-format-truncation: YES 00:05:47.832 Message: lib/log: Defining dependency "log" 00:05:47.832 Message: lib/kvargs: Defining dependency "kvargs" 00:05:47.832 Message: lib/telemetry: Defining dependency "telemetry" 00:05:47.832 Checking for function "getentropy" : NO 00:05:47.832 Message: lib/eal: Defining dependency "eal" 00:05:47.832 Message: lib/ring: Defining dependency "ring" 00:05:47.832 Message: lib/rcu: Defining dependency "rcu" 00:05:47.832 Message: lib/mempool: Defining dependency "mempool" 00:05:47.832 Message: lib/mbuf: Defining dependency "mbuf" 00:05:47.832 Fetching value of define "__PCLMUL__" : 1 (cached) 00:05:47.832 Fetching value of define "__AVX512F__" : 1 (cached) 00:05:47.832 Fetching value of define "__AVX512BW__" : 1 (cached) 00:05:47.832 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:05:47.832 Fetching value of define "__AVX512VL__" : 1 (cached) 00:05:47.832 Fetching value of define "__VPCLMULQDQ__" : 1 (cached) 00:05:47.832 Compiler for C supports arguments -mpclmul: YES 00:05:47.832 Compiler for C supports arguments -maes: YES 00:05:47.832 Compiler for C supports arguments -mavx512f: YES (cached) 00:05:47.832 Compiler for C supports arguments -mavx512bw: YES 00:05:47.832 Compiler for C supports arguments -mavx512dq: YES 00:05:47.832 Compiler for C supports arguments -mavx512vl: YES 00:05:47.833 Compiler for C supports arguments -mvpclmulqdq: YES 00:05:47.833 Compiler for C supports arguments -mavx2: YES 00:05:47.833 Compiler for C supports arguments -mavx: YES 00:05:47.833 Message: lib/net: Defining dependency "net" 00:05:47.833 Message: lib/meter: Defining dependency "meter" 00:05:47.833 Message: lib/ethdev: Defining dependency "ethdev" 00:05:47.833 Message: lib/pci: Defining dependency "pci" 00:05:47.833 Message: lib/cmdline: Defining dependency "cmdline" 00:05:47.833 Message: lib/hash: Defining dependency "hash" 00:05:47.833 Message: lib/timer: Defining dependency "timer" 00:05:47.833 Message: lib/compressdev: Defining dependency "compressdev" 00:05:47.833 Message: lib/cryptodev: Defining dependency "cryptodev" 00:05:47.833 Message: lib/dmadev: Defining dependency "dmadev" 00:05:47.833 Compiler for C supports arguments -Wno-cast-qual: YES 00:05:47.833 Message: lib/power: Defining dependency "power" 00:05:47.833 Message: lib/reorder: Defining dependency "reorder" 00:05:47.833 Message: lib/security: Defining dependency "security" 00:05:47.833 Has header "linux/userfaultfd.h" : YES 00:05:47.833 Has header "linux/vduse.h" : YES 00:05:47.833 Message: lib/vhost: Defining dependency "vhost" 00:05:47.833 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:05:47.833 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:05:47.833 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:05:47.833 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:05:47.833 Message: Disabling raw/* drivers: missing internal dependency "rawdev" 00:05:47.833 Message: Disabling regex/* drivers: missing internal dependency "regexdev" 00:05:47.833 Message: Disabling ml/* drivers: missing internal dependency "mldev" 00:05:47.833 Message: Disabling event/* drivers: missing internal dependency "eventdev" 00:05:47.833 Message: Disabling baseband/* drivers: missing internal dependency "bbdev" 00:05:47.833 Message: Disabling gpu/* drivers: missing internal dependency "gpudev" 00:05:47.833 Program doxygen found: YES (/usr/local/bin/doxygen) 00:05:47.833 Configuring doxy-api-html.conf using configuration 00:05:47.833 Configuring doxy-api-man.conf using configuration 00:05:47.833 Program mandb found: YES (/usr/bin/mandb) 00:05:47.833 Program sphinx-build found: NO 00:05:47.833 Configuring rte_build_config.h using configuration 00:05:47.833 Message: 00:05:47.833 ================= 00:05:47.833 Applications Enabled 00:05:47.833 ================= 00:05:47.833 00:05:47.833 apps: 00:05:47.833 00:05:47.833 00:05:47.833 Message: 00:05:47.833 ================= 00:05:47.833 Libraries Enabled 00:05:47.833 ================= 00:05:47.833 00:05:47.833 libs: 00:05:47.833 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:05:47.833 net, meter, ethdev, pci, cmdline, hash, timer, compressdev, 00:05:47.833 cryptodev, dmadev, power, reorder, security, vhost, 00:05:47.833 00:05:47.833 Message: 00:05:47.833 =============== 00:05:47.833 Drivers Enabled 00:05:47.833 =============== 00:05:47.833 00:05:47.833 common: 00:05:47.833 00:05:47.833 bus: 00:05:47.833 pci, vdev, 00:05:47.833 mempool: 00:05:47.833 ring, 00:05:47.833 dma: 00:05:47.833 00:05:47.833 net: 00:05:47.833 00:05:47.833 crypto: 00:05:47.833 00:05:47.833 compress: 00:05:47.833 00:05:47.833 vdpa: 00:05:47.833 00:05:47.833 00:05:47.833 Message: 00:05:47.833 ================= 00:05:47.833 Content Skipped 00:05:47.833 ================= 00:05:47.833 00:05:47.833 apps: 00:05:47.833 dumpcap: explicitly disabled via build config 00:05:47.833 graph: explicitly disabled via build config 00:05:47.833 pdump: explicitly disabled via build config 00:05:47.833 proc-info: explicitly disabled via build config 00:05:47.833 test-acl: explicitly disabled via build config 00:05:47.833 test-bbdev: explicitly disabled via build config 00:05:47.833 test-cmdline: explicitly disabled via build config 00:05:47.833 test-compress-perf: explicitly disabled via build config 00:05:47.833 test-crypto-perf: explicitly disabled via build config 00:05:47.833 test-dma-perf: explicitly disabled via build config 00:05:47.833 test-eventdev: explicitly disabled via build config 00:05:47.833 test-fib: explicitly disabled via build config 00:05:47.833 test-flow-perf: explicitly disabled via build config 00:05:47.833 test-gpudev: explicitly disabled via build config 00:05:47.833 test-mldev: explicitly disabled via build config 00:05:47.833 test-pipeline: explicitly disabled via build config 00:05:47.833 test-pmd: explicitly disabled via build config 00:05:47.833 test-regex: explicitly disabled via build config 00:05:47.833 test-sad: explicitly disabled via build config 00:05:47.833 test-security-perf: explicitly disabled via build config 00:05:47.833 00:05:47.833 libs: 00:05:47.833 argparse: explicitly disabled via build config 00:05:47.833 metrics: explicitly disabled via build config 00:05:47.833 acl: explicitly disabled via build config 00:05:47.833 bbdev: explicitly disabled via build config 00:05:47.833 bitratestats: explicitly disabled via build config 00:05:47.833 bpf: explicitly disabled via build config 00:05:47.833 cfgfile: explicitly disabled via build config 00:05:47.833 distributor: explicitly disabled via build config 00:05:47.833 efd: explicitly disabled via build config 00:05:47.833 eventdev: explicitly disabled via build config 00:05:47.833 dispatcher: explicitly disabled via build config 00:05:47.833 gpudev: explicitly disabled via build config 00:05:47.833 gro: explicitly disabled via build config 00:05:47.833 gso: explicitly disabled via build config 00:05:47.833 ip_frag: explicitly disabled via build config 00:05:47.833 jobstats: explicitly disabled via build config 00:05:47.833 latencystats: explicitly disabled via build config 00:05:47.833 lpm: explicitly disabled via build config 00:05:47.833 member: explicitly disabled via build config 00:05:47.833 pcapng: explicitly disabled via build config 00:05:47.833 rawdev: explicitly disabled via build config 00:05:47.833 regexdev: explicitly disabled via build config 00:05:47.833 mldev: explicitly disabled via build config 00:05:47.833 rib: explicitly disabled via build config 00:05:47.833 sched: explicitly disabled via build config 00:05:47.833 stack: explicitly disabled via build config 00:05:47.833 ipsec: explicitly disabled via build config 00:05:47.833 pdcp: explicitly disabled via build config 00:05:47.833 fib: explicitly disabled via build config 00:05:47.833 port: explicitly disabled via build config 00:05:47.833 pdump: explicitly disabled via build config 00:05:47.833 table: explicitly disabled via build config 00:05:47.833 pipeline: explicitly disabled via build config 00:05:47.833 graph: explicitly disabled via build config 00:05:47.833 node: explicitly disabled via build config 00:05:47.833 00:05:47.833 drivers: 00:05:47.833 common/cpt: not in enabled drivers build config 00:05:47.833 common/dpaax: not in enabled drivers build config 00:05:47.833 common/iavf: not in enabled drivers build config 00:05:47.833 common/idpf: not in enabled drivers build config 00:05:47.833 common/ionic: not in enabled drivers build config 00:05:47.833 common/mvep: not in enabled drivers build config 00:05:47.833 common/octeontx: not in enabled drivers build config 00:05:47.833 bus/auxiliary: not in enabled drivers build config 00:05:47.833 bus/cdx: not in enabled drivers build config 00:05:47.833 bus/dpaa: not in enabled drivers build config 00:05:47.833 bus/fslmc: not in enabled drivers build config 00:05:47.833 bus/ifpga: not in enabled drivers build config 00:05:47.833 bus/platform: not in enabled drivers build config 00:05:47.833 bus/uacce: not in enabled drivers build config 00:05:47.833 bus/vmbus: not in enabled drivers build config 00:05:47.833 common/cnxk: not in enabled drivers build config 00:05:47.833 common/mlx5: not in enabled drivers build config 00:05:47.833 common/nfp: not in enabled drivers build config 00:05:47.833 common/nitrox: not in enabled drivers build config 00:05:47.833 common/qat: not in enabled drivers build config 00:05:47.833 common/sfc_efx: not in enabled drivers build config 00:05:47.833 mempool/bucket: not in enabled drivers build config 00:05:47.833 mempool/cnxk: not in enabled drivers build config 00:05:47.833 mempool/dpaa: not in enabled drivers build config 00:05:47.833 mempool/dpaa2: not in enabled drivers build config 00:05:47.833 mempool/octeontx: not in enabled drivers build config 00:05:47.833 mempool/stack: not in enabled drivers build config 00:05:47.833 dma/cnxk: not in enabled drivers build config 00:05:47.833 dma/dpaa: not in enabled drivers build config 00:05:47.833 dma/dpaa2: not in enabled drivers build config 00:05:47.833 dma/hisilicon: not in enabled drivers build config 00:05:47.833 dma/idxd: not in enabled drivers build config 00:05:47.833 dma/ioat: not in enabled drivers build config 00:05:47.833 dma/skeleton: not in enabled drivers build config 00:05:47.833 net/af_packet: not in enabled drivers build config 00:05:47.833 net/af_xdp: not in enabled drivers build config 00:05:47.833 net/ark: not in enabled drivers build config 00:05:47.833 net/atlantic: not in enabled drivers build config 00:05:47.833 net/avp: not in enabled drivers build config 00:05:47.833 net/axgbe: not in enabled drivers build config 00:05:47.833 net/bnx2x: not in enabled drivers build config 00:05:47.833 net/bnxt: not in enabled drivers build config 00:05:47.833 net/bonding: not in enabled drivers build config 00:05:47.833 net/cnxk: not in enabled drivers build config 00:05:47.833 net/cpfl: not in enabled drivers build config 00:05:47.833 net/cxgbe: not in enabled drivers build config 00:05:47.833 net/dpaa: not in enabled drivers build config 00:05:47.833 net/dpaa2: not in enabled drivers build config 00:05:47.833 net/e1000: not in enabled drivers build config 00:05:47.833 net/ena: not in enabled drivers build config 00:05:47.833 net/enetc: not in enabled drivers build config 00:05:47.833 net/enetfec: not in enabled drivers build config 00:05:47.833 net/enic: not in enabled drivers build config 00:05:47.833 net/failsafe: not in enabled drivers build config 00:05:47.833 net/fm10k: not in enabled drivers build config 00:05:47.833 net/gve: not in enabled drivers build config 00:05:47.833 net/hinic: not in enabled drivers build config 00:05:47.833 net/hns3: not in enabled drivers build config 00:05:47.833 net/i40e: not in enabled drivers build config 00:05:47.833 net/iavf: not in enabled drivers build config 00:05:47.833 net/ice: not in enabled drivers build config 00:05:47.833 net/idpf: not in enabled drivers build config 00:05:47.833 net/igc: not in enabled drivers build config 00:05:47.833 net/ionic: not in enabled drivers build config 00:05:47.834 net/ipn3ke: not in enabled drivers build config 00:05:47.834 net/ixgbe: not in enabled drivers build config 00:05:47.834 net/mana: not in enabled drivers build config 00:05:47.834 net/memif: not in enabled drivers build config 00:05:47.834 net/mlx4: not in enabled drivers build config 00:05:47.834 net/mlx5: not in enabled drivers build config 00:05:47.834 net/mvneta: not in enabled drivers build config 00:05:47.834 net/mvpp2: not in enabled drivers build config 00:05:47.834 net/netvsc: not in enabled drivers build config 00:05:47.834 net/nfb: not in enabled drivers build config 00:05:47.834 net/nfp: not in enabled drivers build config 00:05:47.834 net/ngbe: not in enabled drivers build config 00:05:47.834 net/null: not in enabled drivers build config 00:05:47.834 net/octeontx: not in enabled drivers build config 00:05:47.834 net/octeon_ep: not in enabled drivers build config 00:05:47.834 net/pcap: not in enabled drivers build config 00:05:47.834 net/pfe: not in enabled drivers build config 00:05:47.834 net/qede: not in enabled drivers build config 00:05:47.834 net/ring: not in enabled drivers build config 00:05:47.834 net/sfc: not in enabled drivers build config 00:05:47.834 net/softnic: not in enabled drivers build config 00:05:47.834 net/tap: not in enabled drivers build config 00:05:47.834 net/thunderx: not in enabled drivers build config 00:05:47.834 net/txgbe: not in enabled drivers build config 00:05:47.834 net/vdev_netvsc: not in enabled drivers build config 00:05:47.834 net/vhost: not in enabled drivers build config 00:05:47.834 net/virtio: not in enabled drivers build config 00:05:47.834 net/vmxnet3: not in enabled drivers build config 00:05:47.834 raw/*: missing internal dependency, "rawdev" 00:05:47.834 crypto/armv8: not in enabled drivers build config 00:05:47.834 crypto/bcmfs: not in enabled drivers build config 00:05:47.834 crypto/caam_jr: not in enabled drivers build config 00:05:47.834 crypto/ccp: not in enabled drivers build config 00:05:47.834 crypto/cnxk: not in enabled drivers build config 00:05:47.834 crypto/dpaa_sec: not in enabled drivers build config 00:05:47.834 crypto/dpaa2_sec: not in enabled drivers build config 00:05:47.834 crypto/ipsec_mb: not in enabled drivers build config 00:05:47.834 crypto/mlx5: not in enabled drivers build config 00:05:47.834 crypto/mvsam: not in enabled drivers build config 00:05:47.834 crypto/nitrox: not in enabled drivers build config 00:05:47.834 crypto/null: not in enabled drivers build config 00:05:47.834 crypto/octeontx: not in enabled drivers build config 00:05:47.834 crypto/openssl: not in enabled drivers build config 00:05:47.834 crypto/scheduler: not in enabled drivers build config 00:05:47.834 crypto/uadk: not in enabled drivers build config 00:05:47.834 crypto/virtio: not in enabled drivers build config 00:05:47.834 compress/isal: not in enabled drivers build config 00:05:47.834 compress/mlx5: not in enabled drivers build config 00:05:47.834 compress/nitrox: not in enabled drivers build config 00:05:47.834 compress/octeontx: not in enabled drivers build config 00:05:47.834 compress/zlib: not in enabled drivers build config 00:05:47.834 regex/*: missing internal dependency, "regexdev" 00:05:47.834 ml/*: missing internal dependency, "mldev" 00:05:47.834 vdpa/ifc: not in enabled drivers build config 00:05:47.834 vdpa/mlx5: not in enabled drivers build config 00:05:47.834 vdpa/nfp: not in enabled drivers build config 00:05:47.834 vdpa/sfc: not in enabled drivers build config 00:05:47.834 event/*: missing internal dependency, "eventdev" 00:05:47.834 baseband/*: missing internal dependency, "bbdev" 00:05:47.834 gpu/*: missing internal dependency, "gpudev" 00:05:47.834 00:05:47.834 00:05:47.834 Build targets in project: 84 00:05:47.834 00:05:47.834 DPDK 24.03.0 00:05:47.834 00:05:47.834 User defined options 00:05:47.834 buildtype : debug 00:05:47.834 default_library : shared 00:05:47.834 libdir : lib 00:05:47.834 prefix : /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:05:47.834 c_args : -Wno-stringop-overflow -fcommon -Wno-stringop-overread -Wno-array-bounds -fPIC -Werror 00:05:47.834 c_link_args : 00:05:47.834 cpu_instruction_set: native 00:05:47.834 disable_apps : test-fib,test-sad,test,test-regex,test-security-perf,test-bbdev,dumpcap,test-crypto-perf,test-flow-perf,test-gpudev,test-cmdline,test-dma-perf,test-eventdev,test-pipeline,test-acl,proc-info,test-compress-perf,graph,test-pmd,test-mldev,pdump 00:05:47.834 disable_libs : bbdev,argparse,latencystats,member,gpudev,mldev,pipeline,lpm,efd,regexdev,sched,node,dispatcher,table,bpf,port,gro,fib,cfgfile,ip_frag,gso,rawdev,ipsec,pdcp,rib,acl,metrics,graph,pcapng,jobstats,eventdev,stack,bitratestats,distributor,pdump 00:05:47.834 enable_docs : false 00:05:47.834 enable_drivers : bus,bus/pci,bus/vdev,mempool/ring 00:05:47.834 enable_kmods : false 00:05:47.834 max_lcores : 128 00:05:47.834 tests : false 00:05:47.834 00:05:47.834 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:05:47.834 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp' 00:05:47.834 [1/267] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:05:47.834 [2/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:05:47.834 [3/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:05:47.834 [4/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:05:47.834 [5/267] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:05:47.834 [6/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:05:47.834 [7/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:05:47.834 [8/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:05:47.834 [9/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:05:47.834 [10/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:05:47.834 [11/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:05:47.834 [12/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:05:47.834 [13/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:05:47.834 [14/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:05:47.834 [15/267] Linking static target lib/librte_kvargs.a 00:05:47.834 [16/267] Compiling C object lib/librte_log.a.p/log_log.c.o 00:05:48.093 [17/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:05:48.093 [18/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:05:48.093 [19/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:05:48.093 [20/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:05:48.093 [21/267] Linking static target lib/librte_log.a 00:05:48.093 [22/267] Compiling C object lib/librte_hash.a.p/hash_rte_hash_crc.c.o 00:05:48.093 [23/267] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:05:48.093 [24/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:05:48.093 [25/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:05:48.093 [26/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:05:48.093 [27/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:05:48.093 [28/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:05:48.093 [29/267] Linking static target lib/librte_pci.a 00:05:48.093 [30/267] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:05:48.093 [31/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:05:48.093 [32/267] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:05:48.093 [33/267] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:05:48.093 [34/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:05:48.093 [35/267] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:05:48.093 [36/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:05:48.093 [37/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:05:48.093 [38/267] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:05:48.355 [39/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:05:48.355 [40/267] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:05:48.355 [41/267] Compiling C object lib/librte_hash.a.p/hash_rte_thash_gfni.c.o 00:05:48.355 [42/267] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:05:48.355 [43/267] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:05:48.355 [44/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:05:48.355 [45/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:05:48.355 [46/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:05:48.355 [47/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:05:48.355 [48/267] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:05:48.355 [49/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:05:48.355 [50/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:05:48.355 [51/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:05:48.355 [52/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:05:48.355 [53/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:05:48.355 [54/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:05:48.355 [55/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:05:48.355 [56/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:05:48.355 [57/267] Compiling C object lib/librte_net.a.p/net_net_crc_avx512.c.o 00:05:48.355 [58/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:05:48.355 [59/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:05:48.355 [60/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:05:48.355 [61/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:05:48.355 [62/267] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:05:48.355 [63/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:05:48.355 [64/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:05:48.355 [65/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:05:48.355 [66/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:05:48.355 [67/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:05:48.355 [68/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:05:48.355 [69/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:05:48.355 [70/267] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:05:48.355 [71/267] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:05:48.355 [72/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:05:48.355 [73/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:05:48.355 [74/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:05:48.355 [75/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:05:48.355 [76/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:05:48.355 [77/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:05:48.355 [78/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:05:48.355 [79/267] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:05:48.355 [80/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:05:48.355 [81/267] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:05:48.355 [82/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:05:48.355 [83/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:05:48.355 [84/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:05:48.355 [85/267] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:05:48.355 [86/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:05:48.355 [87/267] Linking static target lib/librte_telemetry.a 00:05:48.355 [88/267] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:05:48.356 [89/267] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:05:48.356 [90/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:05:48.356 [91/267] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:05:48.356 [92/267] Linking static target lib/librte_meter.a 00:05:48.356 [93/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:05:48.356 [94/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:05:48.616 [95/267] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:05:48.616 [96/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:05:48.616 [97/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:05:48.616 [98/267] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:05:48.616 [99/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:05:48.616 [100/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:05:48.616 [101/267] Linking static target lib/librte_timer.a 00:05:48.616 [102/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:05:48.616 [103/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:05:48.616 [104/267] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:05:48.616 [105/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:05:48.616 [106/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:05:48.616 [107/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:05:48.616 [108/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:05:48.616 [109/267] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:05:48.616 [110/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:05:48.616 [111/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:05:48.616 [112/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:05:48.616 [113/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:05:48.616 [114/267] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:05:48.616 [115/267] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:05:48.616 [116/267] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:05:48.616 [117/267] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:05:48.616 [118/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:05:48.616 [119/267] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:05:48.616 [120/267] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:05:48.616 [121/267] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:05:48.616 [122/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:05:48.616 [123/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_linux_ethtool.c.o 00:05:48.616 [124/267] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:05:48.616 [125/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:05:48.616 [126/267] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:05:48.616 [127/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:05:48.616 [128/267] Linking static target lib/librte_cmdline.a 00:05:48.616 [129/267] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:05:48.616 [130/267] Linking static target lib/librte_ring.a 00:05:48.616 [131/267] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:05:48.617 [132/267] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:05:48.617 [133/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:05:48.617 [134/267] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:05:48.617 [135/267] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:05:48.617 [136/267] Linking static target drivers/libtmp_rte_bus_vdev.a 00:05:48.617 [137/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:05:48.617 [138/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:05:48.617 [139/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:05:48.617 [140/267] Linking static target lib/librte_rcu.a 00:05:48.617 [141/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:05:48.617 [142/267] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:05:48.617 [143/267] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:05:48.617 [144/267] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:05:48.617 [145/267] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:05:48.617 [146/267] Linking static target lib/librte_net.a 00:05:48.617 [147/267] Linking static target lib/librte_compressdev.a 00:05:48.617 [148/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:05:48.617 [149/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:05:48.617 [150/267] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:05:48.617 [151/267] Linking static target lib/librte_dmadev.a 00:05:48.617 [152/267] Linking static target lib/librte_mempool.a 00:05:48.617 [153/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:05:48.617 [154/267] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:05:48.617 [155/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:05:48.617 [156/267] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:05:48.617 [157/267] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:05:48.617 [158/267] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:05:48.617 [159/267] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:05:48.617 [160/267] Linking target lib/librte_log.so.24.1 00:05:48.617 [161/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:05:48.617 [162/267] Linking static target lib/librte_reorder.a 00:05:48.617 [163/267] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:05:48.617 [164/267] Linking static target lib/librte_power.a 00:05:48.617 [165/267] Linking static target lib/librte_mbuf.a 00:05:48.617 [166/267] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:05:48.617 [167/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:05:48.617 [168/267] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:05:48.617 [169/267] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:05:48.617 [170/267] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:05:48.617 [171/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:05:48.617 [172/267] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:05:48.617 [173/267] Linking static target drivers/libtmp_rte_bus_pci.a 00:05:48.617 [174/267] Linking static target drivers/libtmp_rte_mempool_ring.a 00:05:48.617 [175/267] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:05:48.617 [176/267] Linking static target lib/librte_security.a 00:05:48.617 [177/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:05:48.617 [178/267] Linking static target lib/librte_eal.a 00:05:48.617 [179/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:05:48.617 [180/267] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:05:48.877 [181/267] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:05:48.877 [182/267] Generating symbol file lib/librte_log.so.24.1.p/librte_log.so.24.1.symbols 00:05:48.877 [183/267] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:05:48.877 [184/267] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:05:48.877 [185/267] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:05:48.877 [186/267] Compiling C object drivers/librte_bus_vdev.so.24.1.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:05:48.877 [187/267] Linking static target drivers/librte_bus_vdev.a 00:05:48.877 [188/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:05:48.877 [189/267] Linking target lib/librte_kvargs.so.24.1 00:05:48.877 [190/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:05:48.877 [191/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:05:48.877 [192/267] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:05:48.877 [193/267] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:05:48.877 [194/267] Linking static target lib/librte_hash.a 00:05:48.877 [195/267] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:05:48.877 [196/267] Compiling C object drivers/librte_mempool_ring.so.24.1.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:05:48.877 [197/267] Compiling C object drivers/librte_bus_pci.so.24.1.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:05:48.877 [198/267] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:05:48.877 [199/267] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:05:48.877 [200/267] Linking static target drivers/librte_mempool_ring.a 00:05:48.877 [201/267] Linking static target drivers/librte_bus_pci.a 00:05:48.877 [202/267] Generating symbol file lib/librte_kvargs.so.24.1.p/librte_kvargs.so.24.1.symbols 00:05:48.877 [203/267] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:05:48.877 [204/267] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:05:48.877 [205/267] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:05:48.877 [206/267] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.137 [207/267] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.137 [208/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:05:49.137 [209/267] Linking target lib/librte_telemetry.so.24.1 00:05:49.137 [210/267] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:05:49.137 [211/267] Linking static target lib/librte_cryptodev.a 00:05:49.137 [212/267] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.137 [213/267] Generating symbol file lib/librte_telemetry.so.24.1.p/librte_telemetry.so.24.1.symbols 00:05:49.137 [214/267] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.398 [215/267] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:05:49.398 [216/267] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.398 [217/267] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.398 [218/267] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.660 [219/267] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.660 [220/267] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:05:49.660 [221/267] Linking static target lib/librte_ethdev.a 00:05:49.660 [222/267] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.660 [223/267] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.660 [224/267] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.920 [225/267] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:05:49.920 [226/267] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:05:50.491 [227/267] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:05:50.752 [228/267] Linking static target lib/librte_vhost.a 00:05:51.322 [229/267] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:05:52.706 [230/267] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:05:59.335 [231/267] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:06:00.277 [232/267] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:06:00.277 [233/267] Linking target lib/librte_eal.so.24.1 00:06:00.277 [234/267] Generating symbol file lib/librte_eal.so.24.1.p/librte_eal.so.24.1.symbols 00:06:00.277 [235/267] Linking target lib/librte_pci.so.24.1 00:06:00.277 [236/267] Linking target drivers/librte_bus_vdev.so.24.1 00:06:00.277 [237/267] Linking target lib/librte_ring.so.24.1 00:06:00.277 [238/267] Linking target lib/librte_timer.so.24.1 00:06:00.277 [239/267] Linking target lib/librte_meter.so.24.1 00:06:00.277 [240/267] Linking target lib/librte_dmadev.so.24.1 00:06:00.277 [241/267] Generating symbol file lib/librte_pci.so.24.1.p/librte_pci.so.24.1.symbols 00:06:00.277 [242/267] Generating symbol file lib/librte_dmadev.so.24.1.p/librte_dmadev.so.24.1.symbols 00:06:00.537 [243/267] Generating symbol file lib/librte_meter.so.24.1.p/librte_meter.so.24.1.symbols 00:06:00.537 [244/267] Generating symbol file lib/librte_timer.so.24.1.p/librte_timer.so.24.1.symbols 00:06:00.537 [245/267] Generating symbol file lib/librte_ring.so.24.1.p/librte_ring.so.24.1.symbols 00:06:00.537 [246/267] Linking target drivers/librte_bus_pci.so.24.1 00:06:00.537 [247/267] Linking target lib/librte_rcu.so.24.1 00:06:00.537 [248/267] Linking target lib/librte_mempool.so.24.1 00:06:00.537 [249/267] Generating symbol file lib/librte_mempool.so.24.1.p/librte_mempool.so.24.1.symbols 00:06:00.537 [250/267] Generating symbol file lib/librte_rcu.so.24.1.p/librte_rcu.so.24.1.symbols 00:06:00.537 [251/267] Linking target lib/librte_mbuf.so.24.1 00:06:00.537 [252/267] Linking target drivers/librte_mempool_ring.so.24.1 00:06:00.797 [253/267] Generating symbol file lib/librte_mbuf.so.24.1.p/librte_mbuf.so.24.1.symbols 00:06:00.797 [254/267] Linking target lib/librte_compressdev.so.24.1 00:06:00.797 [255/267] Linking target lib/librte_net.so.24.1 00:06:00.797 [256/267] Linking target lib/librte_reorder.so.24.1 00:06:00.797 [257/267] Linking target lib/librte_cryptodev.so.24.1 00:06:01.057 [258/267] Generating symbol file lib/librte_net.so.24.1.p/librte_net.so.24.1.symbols 00:06:01.057 [259/267] Generating symbol file lib/librte_cryptodev.so.24.1.p/librte_cryptodev.so.24.1.symbols 00:06:01.057 [260/267] Linking target lib/librte_hash.so.24.1 00:06:01.057 [261/267] Linking target lib/librte_cmdline.so.24.1 00:06:01.057 [262/267] Linking target lib/librte_ethdev.so.24.1 00:06:01.057 [263/267] Linking target lib/librte_security.so.24.1 00:06:01.057 [264/267] Generating symbol file lib/librte_hash.so.24.1.p/librte_hash.so.24.1.symbols 00:06:01.057 [265/267] Generating symbol file lib/librte_ethdev.so.24.1.p/librte_ethdev.so.24.1.symbols 00:06:01.319 [266/267] Linking target lib/librte_power.so.24.1 00:06:01.319 [267/267] Linking target lib/librte_vhost.so.24.1 00:06:01.319 INFO: autodetecting backend as ninja 00:06:01.319 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp -j 144 00:06:05.523 CC lib/ut_mock/mock.o 00:06:05.523 CC lib/ut/ut.o 00:06:05.523 CC lib/log/log.o 00:06:05.523 CC lib/log/log_deprecated.o 00:06:05.523 CC lib/log/log_flags.o 00:06:05.523 LIB libspdk_ut_mock.a 00:06:05.523 LIB libspdk_log.a 00:06:05.523 LIB libspdk_ut.a 00:06:05.523 SO libspdk_ut_mock.so.6.0 00:06:05.523 SO libspdk_ut.so.2.0 00:06:05.523 SO libspdk_log.so.7.1 00:06:05.523 SYMLINK libspdk_ut_mock.so 00:06:05.523 SYMLINK libspdk_ut.so 00:06:05.523 SYMLINK libspdk_log.so 00:06:06.096 CC lib/ioat/ioat.o 00:06:06.096 CC lib/util/base64.o 00:06:06.096 CC lib/util/bit_array.o 00:06:06.096 CC lib/util/cpuset.o 00:06:06.096 CC lib/dma/dma.o 00:06:06.096 CC lib/util/crc16.o 00:06:06.096 CXX lib/trace_parser/trace.o 00:06:06.096 CC lib/util/crc32.o 00:06:06.096 CC lib/util/crc32c.o 00:06:06.096 CC lib/util/crc32_ieee.o 00:06:06.096 CC lib/util/crc64.o 00:06:06.096 CC lib/util/dif.o 00:06:06.096 CC lib/util/fd.o 00:06:06.096 CC lib/util/fd_group.o 00:06:06.096 CC lib/util/file.o 00:06:06.096 CC lib/util/hexlify.o 00:06:06.096 CC lib/util/iov.o 00:06:06.096 CC lib/util/math.o 00:06:06.096 CC lib/util/net.o 00:06:06.096 CC lib/util/pipe.o 00:06:06.096 CC lib/util/strerror_tls.o 00:06:06.096 CC lib/util/string.o 00:06:06.096 CC lib/util/uuid.o 00:06:06.096 CC lib/util/xor.o 00:06:06.096 CC lib/util/zipf.o 00:06:06.096 CC lib/util/md5.o 00:06:06.096 CC lib/vfio_user/host/vfio_user.o 00:06:06.096 CC lib/vfio_user/host/vfio_user_pci.o 00:06:06.356 LIB libspdk_dma.a 00:06:06.356 SO libspdk_dma.so.5.0 00:06:06.356 LIB libspdk_ioat.a 00:06:06.356 SO libspdk_ioat.so.7.0 00:06:06.356 SYMLINK libspdk_dma.so 00:06:06.357 SYMLINK libspdk_ioat.so 00:06:06.357 LIB libspdk_vfio_user.a 00:06:06.357 SO libspdk_vfio_user.so.5.0 00:06:06.617 LIB libspdk_util.a 00:06:06.617 SYMLINK libspdk_vfio_user.so 00:06:06.617 SO libspdk_util.so.10.0 00:06:06.617 SYMLINK libspdk_util.so 00:06:06.879 LIB libspdk_trace_parser.a 00:06:06.879 SO libspdk_trace_parser.so.6.0 00:06:06.879 SYMLINK libspdk_trace_parser.so 00:06:07.140 CC lib/conf/conf.o 00:06:07.140 CC lib/idxd/idxd.o 00:06:07.140 CC lib/idxd/idxd_user.o 00:06:07.140 CC lib/json/json_parse.o 00:06:07.140 CC lib/json/json_util.o 00:06:07.140 CC lib/idxd/idxd_kernel.o 00:06:07.140 CC lib/json/json_write.o 00:06:07.140 CC lib/rdma_utils/rdma_utils.o 00:06:07.140 CC lib/rdma_provider/common.o 00:06:07.140 CC lib/vmd/vmd.o 00:06:07.140 CC lib/rdma_provider/rdma_provider_verbs.o 00:06:07.140 CC lib/vmd/led.o 00:06:07.140 CC lib/env_dpdk/env.o 00:06:07.140 CC lib/env_dpdk/memory.o 00:06:07.140 CC lib/env_dpdk/pci.o 00:06:07.140 CC lib/env_dpdk/init.o 00:06:07.140 CC lib/env_dpdk/threads.o 00:06:07.140 CC lib/env_dpdk/pci_ioat.o 00:06:07.140 CC lib/env_dpdk/pci_virtio.o 00:06:07.140 CC lib/env_dpdk/pci_vmd.o 00:06:07.140 CC lib/env_dpdk/pci_idxd.o 00:06:07.140 CC lib/env_dpdk/pci_event.o 00:06:07.140 CC lib/env_dpdk/sigbus_handler.o 00:06:07.140 CC lib/env_dpdk/pci_dpdk.o 00:06:07.140 CC lib/env_dpdk/pci_dpdk_2207.o 00:06:07.140 CC lib/env_dpdk/pci_dpdk_2211.o 00:06:07.402 LIB libspdk_rdma_provider.a 00:06:07.402 LIB libspdk_conf.a 00:06:07.402 SO libspdk_rdma_provider.so.6.0 00:06:07.402 SO libspdk_conf.so.6.0 00:06:07.402 LIB libspdk_rdma_utils.a 00:06:07.402 LIB libspdk_json.a 00:06:07.402 SO libspdk_json.so.6.0 00:06:07.402 SO libspdk_rdma_utils.so.1.0 00:06:07.402 SYMLINK libspdk_rdma_provider.so 00:06:07.402 SYMLINK libspdk_conf.so 00:06:07.402 SYMLINK libspdk_rdma_utils.so 00:06:07.402 SYMLINK libspdk_json.so 00:06:07.664 LIB libspdk_idxd.a 00:06:07.664 LIB libspdk_vmd.a 00:06:07.664 SO libspdk_idxd.so.12.1 00:06:07.664 SO libspdk_vmd.so.6.0 00:06:07.664 SYMLINK libspdk_idxd.so 00:06:07.664 SYMLINK libspdk_vmd.so 00:06:07.925 CC lib/jsonrpc/jsonrpc_server.o 00:06:07.925 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:06:07.925 CC lib/jsonrpc/jsonrpc_client.o 00:06:07.925 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:06:08.186 LIB libspdk_jsonrpc.a 00:06:08.186 SO libspdk_jsonrpc.so.6.0 00:06:08.186 SYMLINK libspdk_jsonrpc.so 00:06:08.447 LIB libspdk_env_dpdk.a 00:06:08.447 SO libspdk_env_dpdk.so.15.1 00:06:08.447 SYMLINK libspdk_env_dpdk.so 00:06:08.708 CC lib/rpc/rpc.o 00:06:08.708 LIB libspdk_rpc.a 00:06:08.708 SO libspdk_rpc.so.6.0 00:06:08.968 SYMLINK libspdk_rpc.so 00:06:09.229 CC lib/trace/trace.o 00:06:09.229 CC lib/notify/notify.o 00:06:09.229 CC lib/trace/trace_flags.o 00:06:09.229 CC lib/notify/notify_rpc.o 00:06:09.229 CC lib/keyring/keyring.o 00:06:09.229 CC lib/trace/trace_rpc.o 00:06:09.229 CC lib/keyring/keyring_rpc.o 00:06:09.495 LIB libspdk_notify.a 00:06:09.495 SO libspdk_notify.so.6.0 00:06:09.496 LIB libspdk_keyring.a 00:06:09.496 LIB libspdk_trace.a 00:06:09.496 SO libspdk_keyring.so.2.0 00:06:09.496 SYMLINK libspdk_notify.so 00:06:09.496 SO libspdk_trace.so.11.0 00:06:09.758 SYMLINK libspdk_keyring.so 00:06:09.758 SYMLINK libspdk_trace.so 00:06:10.017 CC lib/thread/thread.o 00:06:10.017 CC lib/sock/sock.o 00:06:10.017 CC lib/thread/iobuf.o 00:06:10.017 CC lib/sock/sock_rpc.o 00:06:10.277 LIB libspdk_sock.a 00:06:10.538 SO libspdk_sock.so.10.0 00:06:10.538 SYMLINK libspdk_sock.so 00:06:10.799 CC lib/nvme/nvme_ctrlr_cmd.o 00:06:10.799 CC lib/nvme/nvme_ctrlr.o 00:06:10.799 CC lib/nvme/nvme_fabric.o 00:06:10.799 CC lib/nvme/nvme_ns_cmd.o 00:06:10.799 CC lib/nvme/nvme_ns.o 00:06:10.799 CC lib/nvme/nvme_pcie_common.o 00:06:10.799 CC lib/nvme/nvme_pcie.o 00:06:10.799 CC lib/nvme/nvme_qpair.o 00:06:10.799 CC lib/nvme/nvme.o 00:06:10.799 CC lib/nvme/nvme_quirks.o 00:06:10.799 CC lib/nvme/nvme_transport.o 00:06:10.799 CC lib/nvme/nvme_discovery.o 00:06:10.799 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:06:10.799 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:06:10.799 CC lib/nvme/nvme_tcp.o 00:06:10.799 CC lib/nvme/nvme_opal.o 00:06:10.799 CC lib/nvme/nvme_io_msg.o 00:06:10.799 CC lib/nvme/nvme_poll_group.o 00:06:10.799 CC lib/nvme/nvme_zns.o 00:06:10.799 CC lib/nvme/nvme_stubs.o 00:06:10.799 CC lib/nvme/nvme_auth.o 00:06:10.799 CC lib/nvme/nvme_vfio_user.o 00:06:10.799 CC lib/nvme/nvme_cuse.o 00:06:10.799 CC lib/nvme/nvme_rdma.o 00:06:11.370 LIB libspdk_thread.a 00:06:11.370 SO libspdk_thread.so.11.0 00:06:11.370 SYMLINK libspdk_thread.so 00:06:11.630 CC lib/accel/accel.o 00:06:11.630 CC lib/accel/accel_rpc.o 00:06:11.630 CC lib/accel/accel_sw.o 00:06:11.630 CC lib/fsdev/fsdev_rpc.o 00:06:11.630 CC lib/fsdev/fsdev.o 00:06:11.630 CC lib/init/json_config.o 00:06:11.630 CC lib/fsdev/fsdev_io.o 00:06:11.630 CC lib/init/subsystem.o 00:06:11.630 CC lib/init/subsystem_rpc.o 00:06:11.630 CC lib/blob/blobstore.o 00:06:11.630 CC lib/init/rpc.o 00:06:11.630 CC lib/blob/request.o 00:06:11.630 CC lib/blob/zeroes.o 00:06:11.630 CC lib/blob/blob_bs_dev.o 00:06:11.630 CC lib/virtio/virtio.o 00:06:11.630 CC lib/vfu_tgt/tgt_endpoint.o 00:06:11.630 CC lib/virtio/virtio_vhost_user.o 00:06:11.630 CC lib/virtio/virtio_vfio_user.o 00:06:11.630 CC lib/virtio/virtio_pci.o 00:06:11.630 CC lib/vfu_tgt/tgt_rpc.o 00:06:11.890 LIB libspdk_init.a 00:06:11.890 SO libspdk_init.so.6.0 00:06:12.150 LIB libspdk_vfu_tgt.a 00:06:12.150 SYMLINK libspdk_init.so 00:06:12.150 LIB libspdk_virtio.a 00:06:12.150 SO libspdk_vfu_tgt.so.3.0 00:06:12.150 SO libspdk_virtio.so.7.0 00:06:12.150 SYMLINK libspdk_vfu_tgt.so 00:06:12.150 SYMLINK libspdk_virtio.so 00:06:12.150 LIB libspdk_fsdev.a 00:06:12.410 SO libspdk_fsdev.so.2.0 00:06:12.410 CC lib/event/app.o 00:06:12.410 CC lib/event/reactor.o 00:06:12.410 CC lib/event/log_rpc.o 00:06:12.410 CC lib/event/app_rpc.o 00:06:12.410 CC lib/event/scheduler_static.o 00:06:12.410 SYMLINK libspdk_fsdev.so 00:06:12.670 LIB libspdk_accel.a 00:06:12.670 SO libspdk_accel.so.16.0 00:06:12.670 LIB libspdk_nvme.a 00:06:12.670 CC lib/fuse_dispatcher/fuse_dispatcher.o 00:06:12.670 SYMLINK libspdk_accel.so 00:06:12.670 SO libspdk_nvme.so.14.1 00:06:12.670 LIB libspdk_event.a 00:06:12.932 SO libspdk_event.so.14.0 00:06:12.932 SYMLINK libspdk_event.so 00:06:12.932 SYMLINK libspdk_nvme.so 00:06:13.193 CC lib/bdev/bdev.o 00:06:13.193 CC lib/bdev/bdev_rpc.o 00:06:13.193 CC lib/bdev/bdev_zone.o 00:06:13.193 CC lib/bdev/part.o 00:06:13.193 CC lib/bdev/scsi_nvme.o 00:06:13.193 LIB libspdk_fuse_dispatcher.a 00:06:13.193 SO libspdk_fuse_dispatcher.so.1.0 00:06:13.453 SYMLINK libspdk_fuse_dispatcher.so 00:06:13.453 LIB libspdk_blob.a 00:06:13.714 SO libspdk_blob.so.11.0 00:06:13.714 SYMLINK libspdk_blob.so 00:06:13.974 CC lib/blobfs/blobfs.o 00:06:13.974 CC lib/blobfs/tree.o 00:06:13.974 CC lib/lvol/lvol.o 00:06:14.916 LIB libspdk_blobfs.a 00:06:14.916 SO libspdk_blobfs.so.10.0 00:06:14.916 LIB libspdk_lvol.a 00:06:14.916 SYMLINK libspdk_blobfs.so 00:06:14.916 SO libspdk_lvol.so.10.0 00:06:14.916 SYMLINK libspdk_lvol.so 00:06:15.487 LIB libspdk_bdev.a 00:06:15.487 SO libspdk_bdev.so.17.0 00:06:15.487 SYMLINK libspdk_bdev.so 00:06:16.057 CC lib/nvmf/ctrlr.o 00:06:16.057 CC lib/nvmf/ctrlr_discovery.o 00:06:16.057 CC lib/nvmf/ctrlr_bdev.o 00:06:16.057 CC lib/nvmf/subsystem.o 00:06:16.057 CC lib/nvmf/nvmf.o 00:06:16.057 CC lib/nvmf/nvmf_rpc.o 00:06:16.057 CC lib/nvmf/transport.o 00:06:16.057 CC lib/nvmf/tcp.o 00:06:16.057 CC lib/nvmf/mdns_server.o 00:06:16.057 CC lib/ftl/ftl_core.o 00:06:16.057 CC lib/nvmf/stubs.o 00:06:16.057 CC lib/ftl/ftl_init.o 00:06:16.057 CC lib/nvmf/vfio_user.o 00:06:16.057 CC lib/ftl/ftl_layout.o 00:06:16.057 CC lib/nvmf/rdma.o 00:06:16.057 CC lib/ftl/ftl_debug.o 00:06:16.057 CC lib/ftl/ftl_io.o 00:06:16.057 CC lib/nvmf/auth.o 00:06:16.057 CC lib/ftl/ftl_sb.o 00:06:16.057 CC lib/scsi/dev.o 00:06:16.057 CC lib/ftl/ftl_l2p.o 00:06:16.057 CC lib/scsi/lun.o 00:06:16.057 CC lib/ublk/ublk.o 00:06:16.057 CC lib/scsi/scsi.o 00:06:16.057 CC lib/ftl/ftl_l2p_flat.o 00:06:16.057 CC lib/scsi/scsi_bdev.o 00:06:16.057 CC lib/scsi/port.o 00:06:16.057 CC lib/ublk/ublk_rpc.o 00:06:16.057 CC lib/ftl/ftl_nv_cache.o 00:06:16.057 CC lib/ftl/ftl_band.o 00:06:16.057 CC lib/scsi/scsi_pr.o 00:06:16.057 CC lib/ftl/ftl_band_ops.o 00:06:16.057 CC lib/scsi/scsi_rpc.o 00:06:16.057 CC lib/ftl/ftl_writer.o 00:06:16.057 CC lib/nbd/nbd.o 00:06:16.057 CC lib/scsi/task.o 00:06:16.057 CC lib/ftl/ftl_rq.o 00:06:16.057 CC lib/nbd/nbd_rpc.o 00:06:16.057 CC lib/ftl/ftl_reloc.o 00:06:16.057 CC lib/ftl/ftl_l2p_cache.o 00:06:16.057 CC lib/ftl/ftl_p2l.o 00:06:16.057 CC lib/ftl/ftl_p2l_log.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_startup.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_md.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_misc.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_band.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:06:16.057 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:06:16.057 CC lib/ftl/utils/ftl_conf.o 00:06:16.057 CC lib/ftl/utils/ftl_md.o 00:06:16.057 CC lib/ftl/utils/ftl_mempool.o 00:06:16.057 CC lib/ftl/utils/ftl_bitmap.o 00:06:16.057 CC lib/ftl/utils/ftl_property.o 00:06:16.057 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:06:16.057 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:06:16.057 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:06:16.057 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:06:16.057 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:06:16.057 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:06:16.057 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:06:16.057 CC lib/ftl/upgrade/ftl_sb_v3.o 00:06:16.057 CC lib/ftl/upgrade/ftl_sb_v5.o 00:06:16.057 CC lib/ftl/nvc/ftl_nvc_dev.o 00:06:16.057 CC lib/ftl/nvc/ftl_nvc_bdev_non_vss.o 00:06:16.057 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:06:16.057 CC lib/ftl/base/ftl_base_dev.o 00:06:16.057 CC lib/ftl/nvc/ftl_nvc_bdev_common.o 00:06:16.057 CC lib/ftl/base/ftl_base_bdev.o 00:06:16.057 CC lib/ftl/ftl_trace.o 00:06:16.626 LIB libspdk_nbd.a 00:06:16.626 SO libspdk_nbd.so.7.0 00:06:16.626 LIB libspdk_scsi.a 00:06:16.626 SYMLINK libspdk_nbd.so 00:06:16.626 SO libspdk_scsi.so.9.0 00:06:16.626 SYMLINK libspdk_scsi.so 00:06:16.626 LIB libspdk_ublk.a 00:06:16.626 SO libspdk_ublk.so.3.0 00:06:16.887 SYMLINK libspdk_ublk.so 00:06:16.887 LIB libspdk_ftl.a 00:06:16.887 CC lib/iscsi/conn.o 00:06:16.887 CC lib/iscsi/init_grp.o 00:06:16.887 CC lib/iscsi/iscsi.o 00:06:16.887 CC lib/iscsi/param.o 00:06:16.887 CC lib/vhost/vhost.o 00:06:16.887 CC lib/iscsi/portal_grp.o 00:06:16.887 CC lib/iscsi/tgt_node.o 00:06:16.887 CC lib/vhost/vhost_rpc.o 00:06:16.887 CC lib/iscsi/task.o 00:06:16.887 CC lib/iscsi/iscsi_subsystem.o 00:06:16.887 CC lib/vhost/vhost_scsi.o 00:06:16.887 CC lib/iscsi/iscsi_rpc.o 00:06:16.887 CC lib/vhost/vhost_blk.o 00:06:16.887 CC lib/vhost/rte_vhost_user.o 00:06:17.146 SO libspdk_ftl.so.9.0 00:06:17.407 SYMLINK libspdk_ftl.so 00:06:17.979 LIB libspdk_nvmf.a 00:06:17.979 SO libspdk_nvmf.so.20.0 00:06:17.979 LIB libspdk_vhost.a 00:06:17.979 SO libspdk_vhost.so.8.0 00:06:17.979 SYMLINK libspdk_vhost.so 00:06:18.240 SYMLINK libspdk_nvmf.so 00:06:18.240 LIB libspdk_iscsi.a 00:06:18.240 SO libspdk_iscsi.so.8.0 00:06:18.240 SYMLINK libspdk_iscsi.so 00:06:18.812 CC module/env_dpdk/env_dpdk_rpc.o 00:06:18.812 CC module/vfu_device/vfu_virtio.o 00:06:18.812 CC module/vfu_device/vfu_virtio_blk.o 00:06:18.812 CC module/vfu_device/vfu_virtio_scsi.o 00:06:18.812 CC module/vfu_device/vfu_virtio_rpc.o 00:06:18.812 CC module/vfu_device/vfu_virtio_fs.o 00:06:19.074 LIB libspdk_env_dpdk_rpc.a 00:06:19.074 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:06:19.074 CC module/keyring/linux/keyring_rpc.o 00:06:19.074 CC module/keyring/linux/keyring.o 00:06:19.074 CC module/keyring/file/keyring.o 00:06:19.074 CC module/keyring/file/keyring_rpc.o 00:06:19.074 CC module/sock/posix/posix.o 00:06:19.074 SO libspdk_env_dpdk_rpc.so.6.0 00:06:19.074 CC module/accel/error/accel_error.o 00:06:19.074 CC module/scheduler/dynamic/scheduler_dynamic.o 00:06:19.074 CC module/accel/error/accel_error_rpc.o 00:06:19.074 CC module/accel/ioat/accel_ioat.o 00:06:19.074 CC module/accel/dsa/accel_dsa.o 00:06:19.074 CC module/accel/ioat/accel_ioat_rpc.o 00:06:19.074 CC module/accel/dsa/accel_dsa_rpc.o 00:06:19.074 CC module/fsdev/aio/fsdev_aio.o 00:06:19.074 CC module/blob/bdev/blob_bdev.o 00:06:19.074 CC module/fsdev/aio/fsdev_aio_rpc.o 00:06:19.074 CC module/fsdev/aio/linux_aio_mgr.o 00:06:19.074 CC module/accel/iaa/accel_iaa.o 00:06:19.074 CC module/scheduler/gscheduler/gscheduler.o 00:06:19.074 CC module/accel/iaa/accel_iaa_rpc.o 00:06:19.074 SYMLINK libspdk_env_dpdk_rpc.so 00:06:19.335 LIB libspdk_keyring_linux.a 00:06:19.335 LIB libspdk_scheduler_dpdk_governor.a 00:06:19.335 LIB libspdk_keyring_file.a 00:06:19.336 SO libspdk_scheduler_dpdk_governor.so.4.0 00:06:19.336 SO libspdk_keyring_linux.so.1.0 00:06:19.336 LIB libspdk_scheduler_gscheduler.a 00:06:19.336 LIB libspdk_scheduler_dynamic.a 00:06:19.336 LIB libspdk_accel_error.a 00:06:19.336 SO libspdk_keyring_file.so.2.0 00:06:19.336 LIB libspdk_accel_ioat.a 00:06:19.336 SO libspdk_scheduler_gscheduler.so.4.0 00:06:19.336 LIB libspdk_accel_iaa.a 00:06:19.336 SO libspdk_scheduler_dynamic.so.4.0 00:06:19.336 SO libspdk_accel_error.so.2.0 00:06:19.336 SO libspdk_accel_ioat.so.6.0 00:06:19.336 SYMLINK libspdk_scheduler_dpdk_governor.so 00:06:19.336 SYMLINK libspdk_keyring_linux.so 00:06:19.336 SYMLINK libspdk_keyring_file.so 00:06:19.336 SO libspdk_accel_iaa.so.3.0 00:06:19.336 SYMLINK libspdk_scheduler_gscheduler.so 00:06:19.336 LIB libspdk_accel_dsa.a 00:06:19.336 LIB libspdk_blob_bdev.a 00:06:19.336 SYMLINK libspdk_accel_error.so 00:06:19.336 SYMLINK libspdk_scheduler_dynamic.so 00:06:19.598 SYMLINK libspdk_accel_ioat.so 00:06:19.598 SO libspdk_accel_dsa.so.5.0 00:06:19.598 SO libspdk_blob_bdev.so.11.0 00:06:19.598 SYMLINK libspdk_accel_iaa.so 00:06:19.598 LIB libspdk_vfu_device.a 00:06:19.598 SYMLINK libspdk_accel_dsa.so 00:06:19.598 SYMLINK libspdk_blob_bdev.so 00:06:19.598 SO libspdk_vfu_device.so.3.0 00:06:19.598 SYMLINK libspdk_vfu_device.so 00:06:19.859 LIB libspdk_fsdev_aio.a 00:06:19.859 LIB libspdk_sock_posix.a 00:06:19.859 SO libspdk_fsdev_aio.so.1.0 00:06:19.859 SO libspdk_sock_posix.so.6.0 00:06:19.859 SYMLINK libspdk_fsdev_aio.so 00:06:19.859 SYMLINK libspdk_sock_posix.so 00:06:20.120 CC module/bdev/gpt/vbdev_gpt.o 00:06:20.120 CC module/bdev/gpt/gpt.o 00:06:20.120 CC module/bdev/error/vbdev_error_rpc.o 00:06:20.120 CC module/bdev/error/vbdev_error.o 00:06:20.120 CC module/bdev/nvme/bdev_nvme.o 00:06:20.120 CC module/bdev/delay/vbdev_delay.o 00:06:20.120 CC module/bdev/nvme/bdev_nvme_rpc.o 00:06:20.120 CC module/bdev/nvme/bdev_mdns_client.o 00:06:20.120 CC module/bdev/delay/vbdev_delay_rpc.o 00:06:20.120 CC module/bdev/nvme/nvme_rpc.o 00:06:20.120 CC module/bdev/nvme/vbdev_opal.o 00:06:20.120 CC module/blobfs/bdev/blobfs_bdev.o 00:06:20.120 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:06:20.120 CC module/bdev/nvme/vbdev_opal_rpc.o 00:06:20.120 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:06:20.120 CC module/bdev/null/bdev_null.o 00:06:20.120 CC module/bdev/lvol/vbdev_lvol.o 00:06:20.120 CC module/bdev/null/bdev_null_rpc.o 00:06:20.120 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:06:20.120 CC module/bdev/malloc/bdev_malloc_rpc.o 00:06:20.120 CC module/bdev/malloc/bdev_malloc.o 00:06:20.120 CC module/bdev/iscsi/bdev_iscsi.o 00:06:20.120 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:06:20.120 CC module/bdev/zone_block/vbdev_zone_block.o 00:06:20.120 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:06:20.120 CC module/bdev/virtio/bdev_virtio_scsi.o 00:06:20.120 CC module/bdev/virtio/bdev_virtio_blk.o 00:06:20.120 CC module/bdev/raid/bdev_raid.o 00:06:20.120 CC module/bdev/virtio/bdev_virtio_rpc.o 00:06:20.120 CC module/bdev/passthru/vbdev_passthru.o 00:06:20.120 CC module/bdev/raid/bdev_raid_rpc.o 00:06:20.120 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:06:20.120 CC module/bdev/raid/raid0.o 00:06:20.120 CC module/bdev/raid/bdev_raid_sb.o 00:06:20.120 CC module/bdev/raid/raid1.o 00:06:20.120 CC module/bdev/split/vbdev_split.o 00:06:20.120 CC module/bdev/raid/concat.o 00:06:20.120 CC module/bdev/ftl/bdev_ftl.o 00:06:20.120 CC module/bdev/split/vbdev_split_rpc.o 00:06:20.120 CC module/bdev/aio/bdev_aio_rpc.o 00:06:20.120 CC module/bdev/ftl/bdev_ftl_rpc.o 00:06:20.120 CC module/bdev/aio/bdev_aio.o 00:06:20.381 LIB libspdk_blobfs_bdev.a 00:06:20.381 SO libspdk_blobfs_bdev.so.6.0 00:06:20.381 LIB libspdk_bdev_gpt.a 00:06:20.381 LIB libspdk_bdev_error.a 00:06:20.381 LIB libspdk_bdev_split.a 00:06:20.381 SO libspdk_bdev_gpt.so.6.0 00:06:20.381 LIB libspdk_bdev_null.a 00:06:20.381 SYMLINK libspdk_blobfs_bdev.so 00:06:20.381 SO libspdk_bdev_error.so.6.0 00:06:20.381 SO libspdk_bdev_split.so.6.0 00:06:20.381 SO libspdk_bdev_null.so.6.0 00:06:20.381 LIB libspdk_bdev_passthru.a 00:06:20.381 LIB libspdk_bdev_ftl.a 00:06:20.381 SYMLINK libspdk_bdev_gpt.so 00:06:20.381 LIB libspdk_bdev_zone_block.a 00:06:20.642 LIB libspdk_bdev_delay.a 00:06:20.642 LIB libspdk_bdev_aio.a 00:06:20.642 SYMLINK libspdk_bdev_error.so 00:06:20.642 SO libspdk_bdev_ftl.so.6.0 00:06:20.642 SO libspdk_bdev_passthru.so.6.0 00:06:20.642 LIB libspdk_bdev_malloc.a 00:06:20.642 LIB libspdk_bdev_iscsi.a 00:06:20.642 SO libspdk_bdev_zone_block.so.6.0 00:06:20.642 SYMLINK libspdk_bdev_split.so 00:06:20.642 SYMLINK libspdk_bdev_null.so 00:06:20.642 SO libspdk_bdev_delay.so.6.0 00:06:20.642 SO libspdk_bdev_aio.so.6.0 00:06:20.642 SO libspdk_bdev_malloc.so.6.0 00:06:20.642 SO libspdk_bdev_iscsi.so.6.0 00:06:20.642 SYMLINK libspdk_bdev_passthru.so 00:06:20.642 SYMLINK libspdk_bdev_ftl.so 00:06:20.642 SYMLINK libspdk_bdev_aio.so 00:06:20.642 SYMLINK libspdk_bdev_zone_block.so 00:06:20.642 SYMLINK libspdk_bdev_delay.so 00:06:20.642 SYMLINK libspdk_bdev_malloc.so 00:06:20.642 SYMLINK libspdk_bdev_iscsi.so 00:06:20.642 LIB libspdk_bdev_lvol.a 00:06:20.642 LIB libspdk_bdev_virtio.a 00:06:20.642 SO libspdk_bdev_lvol.so.6.0 00:06:20.642 SO libspdk_bdev_virtio.so.6.0 00:06:20.642 SYMLINK libspdk_bdev_lvol.so 00:06:20.903 SYMLINK libspdk_bdev_virtio.so 00:06:21.165 LIB libspdk_bdev_raid.a 00:06:21.165 SO libspdk_bdev_raid.so.6.0 00:06:21.165 SYMLINK libspdk_bdev_raid.so 00:06:22.552 LIB libspdk_bdev_nvme.a 00:06:22.552 SO libspdk_bdev_nvme.so.7.0 00:06:22.552 SYMLINK libspdk_bdev_nvme.so 00:06:23.125 CC module/event/subsystems/iobuf/iobuf.o 00:06:23.125 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:06:23.125 CC module/event/subsystems/scheduler/scheduler.o 00:06:23.125 CC module/event/subsystems/vmd/vmd.o 00:06:23.125 CC module/event/subsystems/vmd/vmd_rpc.o 00:06:23.125 CC module/event/subsystems/vfu_tgt/vfu_tgt.o 00:06:23.125 CC module/event/subsystems/sock/sock.o 00:06:23.125 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:06:23.125 CC module/event/subsystems/fsdev/fsdev.o 00:06:23.125 CC module/event/subsystems/keyring/keyring.o 00:06:23.387 LIB libspdk_event_keyring.a 00:06:23.387 SO libspdk_event_keyring.so.1.0 00:06:23.387 LIB libspdk_event_fsdev.a 00:06:23.387 LIB libspdk_event_scheduler.a 00:06:23.387 LIB libspdk_event_vhost_blk.a 00:06:23.387 LIB libspdk_event_vfu_tgt.a 00:06:23.387 LIB libspdk_event_iobuf.a 00:06:23.387 LIB libspdk_event_vmd.a 00:06:23.387 LIB libspdk_event_sock.a 00:06:23.387 SO libspdk_event_fsdev.so.1.0 00:06:23.387 SO libspdk_event_vfu_tgt.so.3.0 00:06:23.387 SO libspdk_event_vmd.so.6.0 00:06:23.387 SYMLINK libspdk_event_keyring.so 00:06:23.387 SO libspdk_event_scheduler.so.4.0 00:06:23.387 SO libspdk_event_iobuf.so.3.0 00:06:23.387 SO libspdk_event_sock.so.5.0 00:06:23.387 SO libspdk_event_vhost_blk.so.3.0 00:06:23.387 SYMLINK libspdk_event_fsdev.so 00:06:23.387 SYMLINK libspdk_event_vfu_tgt.so 00:06:23.387 SYMLINK libspdk_event_vhost_blk.so 00:06:23.387 SYMLINK libspdk_event_sock.so 00:06:23.387 SYMLINK libspdk_event_vmd.so 00:06:23.387 SYMLINK libspdk_event_scheduler.so 00:06:23.387 SYMLINK libspdk_event_iobuf.so 00:06:23.958 CC module/event/subsystems/accel/accel.o 00:06:23.958 LIB libspdk_event_accel.a 00:06:23.958 SO libspdk_event_accel.so.6.0 00:06:23.958 SYMLINK libspdk_event_accel.so 00:06:24.528 CC module/event/subsystems/bdev/bdev.o 00:06:24.528 LIB libspdk_event_bdev.a 00:06:24.528 SO libspdk_event_bdev.so.6.0 00:06:24.528 SYMLINK libspdk_event_bdev.so 00:06:25.102 CC module/event/subsystems/scsi/scsi.o 00:06:25.102 CC module/event/subsystems/nbd/nbd.o 00:06:25.102 CC module/event/subsystems/ublk/ublk.o 00:06:25.102 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:06:25.103 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:06:25.103 LIB libspdk_event_nbd.a 00:06:25.103 LIB libspdk_event_scsi.a 00:06:25.103 LIB libspdk_event_ublk.a 00:06:25.103 SO libspdk_event_nbd.so.6.0 00:06:25.103 SO libspdk_event_scsi.so.6.0 00:06:25.103 SO libspdk_event_ublk.so.3.0 00:06:25.435 LIB libspdk_event_nvmf.a 00:06:25.435 SYMLINK libspdk_event_nbd.so 00:06:25.435 SYMLINK libspdk_event_scsi.so 00:06:25.435 SYMLINK libspdk_event_ublk.so 00:06:25.435 SO libspdk_event_nvmf.so.6.0 00:06:25.435 SYMLINK libspdk_event_nvmf.so 00:06:25.697 CC module/event/subsystems/iscsi/iscsi.o 00:06:25.697 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:06:25.697 LIB libspdk_event_iscsi.a 00:06:25.697 LIB libspdk_event_vhost_scsi.a 00:06:25.957 SO libspdk_event_iscsi.so.6.0 00:06:25.957 SO libspdk_event_vhost_scsi.so.3.0 00:06:25.957 SYMLINK libspdk_event_iscsi.so 00:06:25.957 SYMLINK libspdk_event_vhost_scsi.so 00:06:26.217 SO libspdk.so.6.0 00:06:26.217 SYMLINK libspdk.so 00:06:26.480 CC app/trace_record/trace_record.o 00:06:26.480 CC app/spdk_top/spdk_top.o 00:06:26.480 CXX app/trace/trace.o 00:06:26.480 CC app/spdk_lspci/spdk_lspci.o 00:06:26.480 TEST_HEADER include/spdk/accel.h 00:06:26.480 TEST_HEADER include/spdk/accel_module.h 00:06:26.480 CC app/spdk_nvme_discover/discovery_aer.o 00:06:26.480 TEST_HEADER include/spdk/assert.h 00:06:26.480 TEST_HEADER include/spdk/barrier.h 00:06:26.480 CC app/spdk_nvme_identify/identify.o 00:06:26.480 TEST_HEADER include/spdk/base64.h 00:06:26.480 TEST_HEADER include/spdk/bdev.h 00:06:26.480 CC app/spdk_nvme_perf/perf.o 00:06:26.480 TEST_HEADER include/spdk/bdev_module.h 00:06:26.480 TEST_HEADER include/spdk/bdev_zone.h 00:06:26.480 TEST_HEADER include/spdk/bit_pool.h 00:06:26.480 TEST_HEADER include/spdk/bit_array.h 00:06:26.480 TEST_HEADER include/spdk/blob_bdev.h 00:06:26.480 TEST_HEADER include/spdk/blobfs_bdev.h 00:06:26.480 CC test/rpc_client/rpc_client_test.o 00:06:26.480 TEST_HEADER include/spdk/blobfs.h 00:06:26.480 TEST_HEADER include/spdk/blob.h 00:06:26.480 TEST_HEADER include/spdk/conf.h 00:06:26.480 TEST_HEADER include/spdk/config.h 00:06:26.480 TEST_HEADER include/spdk/cpuset.h 00:06:26.480 TEST_HEADER include/spdk/crc16.h 00:06:26.480 TEST_HEADER include/spdk/crc32.h 00:06:26.480 TEST_HEADER include/spdk/dif.h 00:06:26.480 TEST_HEADER include/spdk/crc64.h 00:06:26.480 TEST_HEADER include/spdk/dma.h 00:06:26.480 TEST_HEADER include/spdk/endian.h 00:06:26.480 TEST_HEADER include/spdk/env.h 00:06:26.480 TEST_HEADER include/spdk/env_dpdk.h 00:06:26.480 TEST_HEADER include/spdk/fd_group.h 00:06:26.480 TEST_HEADER include/spdk/event.h 00:06:26.480 TEST_HEADER include/spdk/fd.h 00:06:26.480 TEST_HEADER include/spdk/file.h 00:06:26.480 TEST_HEADER include/spdk/ftl.h 00:06:26.480 TEST_HEADER include/spdk/fsdev_module.h 00:06:26.480 TEST_HEADER include/spdk/fuse_dispatcher.h 00:06:26.480 TEST_HEADER include/spdk/fsdev.h 00:06:26.480 TEST_HEADER include/spdk/hexlify.h 00:06:26.480 TEST_HEADER include/spdk/gpt_spec.h 00:06:26.480 TEST_HEADER include/spdk/histogram_data.h 00:06:26.480 TEST_HEADER include/spdk/idxd.h 00:06:26.480 TEST_HEADER include/spdk/idxd_spec.h 00:06:26.480 TEST_HEADER include/spdk/init.h 00:06:26.480 TEST_HEADER include/spdk/ioat.h 00:06:26.480 TEST_HEADER include/spdk/ioat_spec.h 00:06:26.480 TEST_HEADER include/spdk/iscsi_spec.h 00:06:26.480 TEST_HEADER include/spdk/json.h 00:06:26.480 CC app/nvmf_tgt/nvmf_main.o 00:06:26.480 TEST_HEADER include/spdk/jsonrpc.h 00:06:26.480 TEST_HEADER include/spdk/keyring.h 00:06:26.480 CC examples/interrupt_tgt/interrupt_tgt.o 00:06:26.480 TEST_HEADER include/spdk/keyring_module.h 00:06:26.480 TEST_HEADER include/spdk/lvol.h 00:06:26.480 TEST_HEADER include/spdk/likely.h 00:06:26.480 TEST_HEADER include/spdk/log.h 00:06:26.480 TEST_HEADER include/spdk/md5.h 00:06:26.480 TEST_HEADER include/spdk/mmio.h 00:06:26.480 CC app/spdk_dd/spdk_dd.o 00:06:26.480 TEST_HEADER include/spdk/memory.h 00:06:26.480 CC app/iscsi_tgt/iscsi_tgt.o 00:06:26.480 TEST_HEADER include/spdk/nbd.h 00:06:26.480 TEST_HEADER include/spdk/net.h 00:06:26.480 TEST_HEADER include/spdk/notify.h 00:06:26.480 TEST_HEADER include/spdk/nvme.h 00:06:26.480 TEST_HEADER include/spdk/nvme_intel.h 00:06:26.480 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:06:26.480 TEST_HEADER include/spdk/nvme_ocssd.h 00:06:26.480 TEST_HEADER include/spdk/nvme_spec.h 00:06:26.480 TEST_HEADER include/spdk/nvme_zns.h 00:06:26.480 TEST_HEADER include/spdk/nvmf_cmd.h 00:06:26.480 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:06:26.480 TEST_HEADER include/spdk/nvmf.h 00:06:26.480 TEST_HEADER include/spdk/nvmf_spec.h 00:06:26.480 TEST_HEADER include/spdk/opal.h 00:06:26.480 TEST_HEADER include/spdk/nvmf_transport.h 00:06:26.480 TEST_HEADER include/spdk/opal_spec.h 00:06:26.480 TEST_HEADER include/spdk/pci_ids.h 00:06:26.480 TEST_HEADER include/spdk/queue.h 00:06:26.480 CC app/spdk_tgt/spdk_tgt.o 00:06:26.480 TEST_HEADER include/spdk/pipe.h 00:06:26.480 TEST_HEADER include/spdk/reduce.h 00:06:26.480 TEST_HEADER include/spdk/rpc.h 00:06:26.480 TEST_HEADER include/spdk/scsi.h 00:06:26.480 TEST_HEADER include/spdk/scheduler.h 00:06:26.480 TEST_HEADER include/spdk/scsi_spec.h 00:06:26.480 TEST_HEADER include/spdk/string.h 00:06:26.480 TEST_HEADER include/spdk/sock.h 00:06:26.480 TEST_HEADER include/spdk/stdinc.h 00:06:26.480 TEST_HEADER include/spdk/thread.h 00:06:26.480 TEST_HEADER include/spdk/trace.h 00:06:26.746 TEST_HEADER include/spdk/trace_parser.h 00:06:26.746 TEST_HEADER include/spdk/ublk.h 00:06:26.746 TEST_HEADER include/spdk/util.h 00:06:26.746 TEST_HEADER include/spdk/tree.h 00:06:26.746 TEST_HEADER include/spdk/version.h 00:06:26.746 TEST_HEADER include/spdk/uuid.h 00:06:26.746 TEST_HEADER include/spdk/vfio_user_pci.h 00:06:26.746 TEST_HEADER include/spdk/vfio_user_spec.h 00:06:26.746 TEST_HEADER include/spdk/vmd.h 00:06:26.746 TEST_HEADER include/spdk/vhost.h 00:06:26.746 TEST_HEADER include/spdk/xor.h 00:06:26.746 TEST_HEADER include/spdk/zipf.h 00:06:26.746 CXX test/cpp_headers/accel.o 00:06:26.746 CXX test/cpp_headers/accel_module.o 00:06:26.746 CXX test/cpp_headers/assert.o 00:06:26.746 CXX test/cpp_headers/barrier.o 00:06:26.746 CXX test/cpp_headers/base64.o 00:06:26.746 CXX test/cpp_headers/bit_array.o 00:06:26.746 CXX test/cpp_headers/bdev.o 00:06:26.746 CXX test/cpp_headers/bdev_module.o 00:06:26.746 CXX test/cpp_headers/bdev_zone.o 00:06:26.746 CXX test/cpp_headers/bit_pool.o 00:06:26.746 CXX test/cpp_headers/blob_bdev.o 00:06:26.746 CXX test/cpp_headers/blobfs.o 00:06:26.746 CXX test/cpp_headers/blobfs_bdev.o 00:06:26.746 CXX test/cpp_headers/blob.o 00:06:26.746 CXX test/cpp_headers/conf.o 00:06:26.746 CXX test/cpp_headers/config.o 00:06:26.746 CXX test/cpp_headers/cpuset.o 00:06:26.746 CXX test/cpp_headers/crc16.o 00:06:26.746 CXX test/cpp_headers/crc32.o 00:06:26.746 CXX test/cpp_headers/dif.o 00:06:26.746 CXX test/cpp_headers/crc64.o 00:06:26.746 CXX test/cpp_headers/dma.o 00:06:26.746 CXX test/cpp_headers/endian.o 00:06:26.746 CXX test/cpp_headers/env.o 00:06:26.746 CXX test/cpp_headers/env_dpdk.o 00:06:26.746 CXX test/cpp_headers/event.o 00:06:26.746 CXX test/cpp_headers/fd_group.o 00:06:26.746 CXX test/cpp_headers/fd.o 00:06:26.746 CXX test/cpp_headers/fsdev_module.o 00:06:26.746 CXX test/cpp_headers/file.o 00:06:26.746 CXX test/cpp_headers/fsdev.o 00:06:26.746 CXX test/cpp_headers/ftl.o 00:06:26.746 CXX test/cpp_headers/fuse_dispatcher.o 00:06:26.746 CXX test/cpp_headers/gpt_spec.o 00:06:26.746 CXX test/cpp_headers/hexlify.o 00:06:26.746 CXX test/cpp_headers/histogram_data.o 00:06:26.746 CXX test/cpp_headers/idxd.o 00:06:26.746 CXX test/cpp_headers/init.o 00:06:26.746 CXX test/cpp_headers/idxd_spec.o 00:06:26.746 CXX test/cpp_headers/ioat.o 00:06:26.746 CXX test/cpp_headers/ioat_spec.o 00:06:26.746 CXX test/cpp_headers/iscsi_spec.o 00:06:26.746 CXX test/cpp_headers/jsonrpc.o 00:06:26.746 CXX test/cpp_headers/json.o 00:06:26.746 CXX test/cpp_headers/keyring_module.o 00:06:26.746 CXX test/cpp_headers/keyring.o 00:06:26.746 CXX test/cpp_headers/likely.o 00:06:26.746 CXX test/cpp_headers/md5.o 00:06:26.746 CXX test/cpp_headers/lvol.o 00:06:26.746 CXX test/cpp_headers/log.o 00:06:26.746 CXX test/cpp_headers/memory.o 00:06:26.746 CXX test/cpp_headers/mmio.o 00:06:26.746 CXX test/cpp_headers/net.o 00:06:26.746 CXX test/cpp_headers/nbd.o 00:06:26.746 CXX test/cpp_headers/notify.o 00:06:26.746 CXX test/cpp_headers/nvme.o 00:06:26.746 CXX test/cpp_headers/nvme_intel.o 00:06:26.746 CXX test/cpp_headers/nvme_ocssd.o 00:06:26.746 CXX test/cpp_headers/nvme_spec.o 00:06:26.746 CXX test/cpp_headers/nvme_ocssd_spec.o 00:06:26.746 CXX test/cpp_headers/nvme_zns.o 00:06:26.746 CXX test/cpp_headers/nvmf.o 00:06:26.746 CXX test/cpp_headers/nvmf_fc_spec.o 00:06:26.746 CXX test/cpp_headers/nvmf_transport.o 00:06:26.746 CXX test/cpp_headers/nvmf_cmd.o 00:06:26.746 CXX test/cpp_headers/nvmf_spec.o 00:06:26.746 CXX test/cpp_headers/pci_ids.o 00:06:26.746 CXX test/cpp_headers/opal.o 00:06:26.746 CXX test/cpp_headers/opal_spec.o 00:06:26.746 CXX test/cpp_headers/pipe.o 00:06:26.746 CC examples/ioat/perf/perf.o 00:06:26.746 CXX test/cpp_headers/queue.o 00:06:26.746 CXX test/cpp_headers/reduce.o 00:06:26.746 CXX test/cpp_headers/scsi.o 00:06:26.746 CXX test/cpp_headers/rpc.o 00:06:26.746 CXX test/cpp_headers/scsi_spec.o 00:06:26.746 CXX test/cpp_headers/scheduler.o 00:06:26.746 CXX test/cpp_headers/sock.o 00:06:26.746 CXX test/cpp_headers/stdinc.o 00:06:26.746 CXX test/cpp_headers/thread.o 00:06:26.746 CXX test/cpp_headers/string.o 00:06:26.746 CXX test/cpp_headers/trace.o 00:06:26.746 CXX test/cpp_headers/ublk.o 00:06:26.746 CXX test/cpp_headers/trace_parser.o 00:06:26.746 CXX test/cpp_headers/tree.o 00:06:26.746 LINK spdk_lspci 00:06:26.746 CXX test/cpp_headers/util.o 00:06:26.746 CC examples/util/zipf/zipf.o 00:06:26.746 CC app/fio/nvme/fio_plugin.o 00:06:26.746 CXX test/cpp_headers/uuid.o 00:06:26.746 CXX test/cpp_headers/version.o 00:06:26.746 CXX test/cpp_headers/vmd.o 00:06:26.746 CXX test/cpp_headers/vfio_user_pci.o 00:06:26.746 CXX test/cpp_headers/vhost.o 00:06:26.746 CXX test/cpp_headers/xor.o 00:06:26.746 CXX test/cpp_headers/vfio_user_spec.o 00:06:26.746 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:06:26.746 CXX test/cpp_headers/zipf.o 00:06:26.746 CC test/app/jsoncat/jsoncat.o 00:06:26.746 CC test/env/memory/memory_ut.o 00:06:26.746 CC test/env/vtophys/vtophys.o 00:06:26.746 CC test/env/pci/pci_ut.o 00:06:26.746 CC examples/ioat/verify/verify.o 00:06:26.746 CC test/app/stub/stub.o 00:06:26.746 CC test/thread/poller_perf/poller_perf.o 00:06:26.746 CC test/app/histogram_perf/histogram_perf.o 00:06:26.746 CC app/fio/bdev/fio_plugin.o 00:06:26.746 CC test/app/bdev_svc/bdev_svc.o 00:06:26.746 LINK rpc_client_test 00:06:26.746 CC test/dma/test_dma/test_dma.o 00:06:27.009 LINK spdk_nvme_discover 00:06:27.009 LINK interrupt_tgt 00:06:27.009 LINK spdk_trace_record 00:06:27.009 LINK nvmf_tgt 00:06:27.009 LINK iscsi_tgt 00:06:27.009 LINK spdk_tgt 00:06:27.269 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:06:27.269 CC test/env/mem_callbacks/mem_callbacks.o 00:06:27.269 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:06:27.269 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:06:27.269 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:06:27.269 LINK histogram_perf 00:06:27.269 LINK zipf 00:06:27.269 LINK env_dpdk_post_init 00:06:27.269 LINK bdev_svc 00:06:27.269 LINK jsoncat 00:06:27.269 LINK spdk_dd 00:06:27.530 LINK vtophys 00:06:27.530 LINK ioat_perf 00:06:27.530 LINK spdk_trace 00:06:27.530 LINK poller_perf 00:06:27.530 LINK stub 00:06:27.530 LINK verify 00:06:27.530 CC examples/vmd/led/led.o 00:06:27.530 CC examples/vmd/lsvmd/lsvmd.o 00:06:27.792 CC examples/idxd/perf/perf.o 00:06:27.792 CC examples/sock/hello_world/hello_sock.o 00:06:27.792 LINK spdk_top 00:06:27.792 LINK pci_ut 00:06:27.792 CC examples/thread/thread/thread_ex.o 00:06:27.792 LINK spdk_nvme 00:06:27.792 LINK spdk_bdev 00:06:27.792 CC app/vhost/vhost.o 00:06:27.792 LINK lsvmd 00:06:27.792 LINK vhost_fuzz 00:06:27.792 LINK spdk_nvme_perf 00:06:27.792 LINK led 00:06:27.792 LINK nvme_fuzz 00:06:27.792 LINK test_dma 00:06:27.792 LINK spdk_nvme_identify 00:06:27.792 LINK mem_callbacks 00:06:27.792 CC test/event/event_perf/event_perf.o 00:06:27.792 CC test/event/reactor/reactor.o 00:06:27.792 CC test/event/reactor_perf/reactor_perf.o 00:06:27.792 CC test/event/app_repeat/app_repeat.o 00:06:28.054 LINK hello_sock 00:06:28.054 CC test/event/scheduler/scheduler.o 00:06:28.054 LINK thread 00:06:28.054 LINK idxd_perf 00:06:28.054 LINK vhost 00:06:28.054 LINK reactor 00:06:28.054 LINK event_perf 00:06:28.054 LINK reactor_perf 00:06:28.054 LINK app_repeat 00:06:28.315 LINK scheduler 00:06:28.315 LINK memory_ut 00:06:28.315 CC test/nvme/simple_copy/simple_copy.o 00:06:28.315 CC test/nvme/aer/aer.o 00:06:28.315 CC test/nvme/e2edp/nvme_dp.o 00:06:28.315 CC test/nvme/boot_partition/boot_partition.o 00:06:28.315 CC test/nvme/sgl/sgl.o 00:06:28.315 CC test/nvme/err_injection/err_injection.o 00:06:28.575 CC test/nvme/doorbell_aers/doorbell_aers.o 00:06:28.575 CC test/nvme/reset/reset.o 00:06:28.575 CC test/nvme/compliance/nvme_compliance.o 00:06:28.575 CC test/nvme/reserve/reserve.o 00:06:28.575 CC test/nvme/connect_stress/connect_stress.o 00:06:28.575 CC test/nvme/fused_ordering/fused_ordering.o 00:06:28.575 CC test/nvme/cuse/cuse.o 00:06:28.575 CC test/nvme/overhead/overhead.o 00:06:28.575 CC test/nvme/startup/startup.o 00:06:28.575 CC test/nvme/fdp/fdp.o 00:06:28.575 CC examples/nvme/cmb_copy/cmb_copy.o 00:06:28.575 CC examples/nvme/arbitration/arbitration.o 00:06:28.575 CC examples/nvme/nvme_manage/nvme_manage.o 00:06:28.575 CC test/accel/dif/dif.o 00:06:28.575 CC examples/nvme/hello_world/hello_world.o 00:06:28.575 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:06:28.575 CC examples/nvme/abort/abort.o 00:06:28.575 CC examples/nvme/reconnect/reconnect.o 00:06:28.575 CC test/blobfs/mkfs/mkfs.o 00:06:28.575 CC examples/nvme/hotplug/hotplug.o 00:06:28.575 CC examples/accel/perf/accel_perf.o 00:06:28.575 CC examples/blob/cli/blobcli.o 00:06:28.575 CC examples/blob/hello_world/hello_blob.o 00:06:28.575 LINK doorbell_aers 00:06:28.575 CC examples/fsdev/hello_world/hello_fsdev.o 00:06:28.575 CC test/lvol/esnap/esnap.o 00:06:28.575 LINK boot_partition 00:06:28.575 LINK err_injection 00:06:28.575 LINK connect_stress 00:06:28.575 LINK startup 00:06:28.575 LINK fused_ordering 00:06:28.575 LINK reserve 00:06:28.575 LINK simple_copy 00:06:28.575 LINK pmr_persistence 00:06:28.575 LINK cmb_copy 00:06:28.835 LINK sgl 00:06:28.835 LINK hello_world 00:06:28.835 LINK reset 00:06:28.835 LINK aer 00:06:28.835 LINK mkfs 00:06:28.835 LINK nvme_dp 00:06:28.835 LINK hotplug 00:06:28.835 LINK overhead 00:06:28.835 LINK iscsi_fuzz 00:06:28.835 LINK nvme_compliance 00:06:28.835 LINK fdp 00:06:28.835 LINK arbitration 00:06:28.835 LINK reconnect 00:06:28.835 LINK abort 00:06:28.835 LINK hello_blob 00:06:28.835 LINK nvme_manage 00:06:28.835 LINK hello_fsdev 00:06:29.096 LINK accel_perf 00:06:29.096 LINK blobcli 00:06:29.096 LINK dif 00:06:29.674 CC examples/bdev/hello_world/hello_bdev.o 00:06:29.674 CC examples/bdev/bdevperf/bdevperf.o 00:06:29.674 LINK cuse 00:06:29.674 CC test/bdev/bdevio/bdevio.o 00:06:29.938 LINK hello_bdev 00:06:30.199 LINK bdevio 00:06:30.523 LINK bdevperf 00:06:31.201 CC examples/nvmf/nvmf/nvmf.o 00:06:31.201 LINK nvmf 00:06:33.118 LINK esnap 00:06:33.692 00:06:33.692 real 0m55.280s 00:06:33.692 user 7m46.256s 00:06:33.692 sys 4m23.722s 00:06:33.692 08:42:23 make -- common/autotest_common.sh@1126 -- $ xtrace_disable 00:06:33.692 08:42:23 make -- common/autotest_common.sh@10 -- $ set +x 00:06:33.692 ************************************ 00:06:33.692 END TEST make 00:06:33.692 ************************************ 00:06:33.692 08:42:23 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:06:33.692 08:42:23 -- pm/common@29 -- $ signal_monitor_resources TERM 00:06:33.692 08:42:23 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:06:33.692 08:42:23 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:06:33.692 08:42:23 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:06:33.692 08:42:23 -- pm/common@44 -- $ pid=1566703 00:06:33.692 08:42:23 -- pm/common@50 -- $ kill -TERM 1566703 00:06:33.692 08:42:23 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:06:33.692 08:42:23 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:06:33.692 08:42:23 -- pm/common@44 -- $ pid=1566704 00:06:33.692 08:42:23 -- pm/common@50 -- $ kill -TERM 1566704 00:06:33.692 08:42:23 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:06:33.692 08:42:23 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:06:33.692 08:42:23 -- pm/common@44 -- $ pid=1566706 00:06:33.692 08:42:23 -- pm/common@50 -- $ kill -TERM 1566706 00:06:33.692 08:42:23 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:06:33.692 08:42:23 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:06:33.692 08:42:23 -- pm/common@44 -- $ pid=1566730 00:06:33.692 08:42:23 -- pm/common@50 -- $ sudo -E kill -TERM 1566730 00:06:33.692 08:42:23 -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:06:33.692 08:42:23 -- common/autotest_common.sh@1689 -- # lcov --version 00:06:33.692 08:42:23 -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:06:33.692 08:42:23 -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:06:33.692 08:42:23 -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:06:33.692 08:42:23 -- scripts/common.sh@333 -- # local ver1 ver1_l 00:06:33.692 08:42:23 -- scripts/common.sh@334 -- # local ver2 ver2_l 00:06:33.692 08:42:23 -- scripts/common.sh@336 -- # IFS=.-: 00:06:33.692 08:42:23 -- scripts/common.sh@336 -- # read -ra ver1 00:06:33.692 08:42:23 -- scripts/common.sh@337 -- # IFS=.-: 00:06:33.692 08:42:23 -- scripts/common.sh@337 -- # read -ra ver2 00:06:33.692 08:42:23 -- scripts/common.sh@338 -- # local 'op=<' 00:06:33.692 08:42:23 -- scripts/common.sh@340 -- # ver1_l=2 00:06:33.692 08:42:23 -- scripts/common.sh@341 -- # ver2_l=1 00:06:33.692 08:42:23 -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:06:33.692 08:42:23 -- scripts/common.sh@344 -- # case "$op" in 00:06:33.692 08:42:23 -- scripts/common.sh@345 -- # : 1 00:06:33.692 08:42:23 -- scripts/common.sh@364 -- # (( v = 0 )) 00:06:33.692 08:42:23 -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:06:33.692 08:42:23 -- scripts/common.sh@365 -- # decimal 1 00:06:33.692 08:42:23 -- scripts/common.sh@353 -- # local d=1 00:06:33.692 08:42:23 -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:06:33.692 08:42:23 -- scripts/common.sh@355 -- # echo 1 00:06:33.692 08:42:23 -- scripts/common.sh@365 -- # ver1[v]=1 00:06:33.692 08:42:23 -- scripts/common.sh@366 -- # decimal 2 00:06:33.692 08:42:23 -- scripts/common.sh@353 -- # local d=2 00:06:33.692 08:42:23 -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:06:33.692 08:42:23 -- scripts/common.sh@355 -- # echo 2 00:06:33.692 08:42:23 -- scripts/common.sh@366 -- # ver2[v]=2 00:06:33.692 08:42:23 -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:06:33.692 08:42:23 -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:06:33.692 08:42:23 -- scripts/common.sh@368 -- # return 0 00:06:33.692 08:42:23 -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:06:33.692 08:42:23 -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:06:33.692 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:33.692 --rc genhtml_branch_coverage=1 00:06:33.692 --rc genhtml_function_coverage=1 00:06:33.692 --rc genhtml_legend=1 00:06:33.692 --rc geninfo_all_blocks=1 00:06:33.692 --rc geninfo_unexecuted_blocks=1 00:06:33.692 00:06:33.692 ' 00:06:33.692 08:42:23 -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:06:33.692 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:33.692 --rc genhtml_branch_coverage=1 00:06:33.692 --rc genhtml_function_coverage=1 00:06:33.692 --rc genhtml_legend=1 00:06:33.692 --rc geninfo_all_blocks=1 00:06:33.692 --rc geninfo_unexecuted_blocks=1 00:06:33.692 00:06:33.692 ' 00:06:33.692 08:42:23 -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:06:33.692 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:33.692 --rc genhtml_branch_coverage=1 00:06:33.692 --rc genhtml_function_coverage=1 00:06:33.692 --rc genhtml_legend=1 00:06:33.692 --rc geninfo_all_blocks=1 00:06:33.692 --rc geninfo_unexecuted_blocks=1 00:06:33.692 00:06:33.692 ' 00:06:33.692 08:42:23 -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:06:33.692 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:06:33.692 --rc genhtml_branch_coverage=1 00:06:33.692 --rc genhtml_function_coverage=1 00:06:33.692 --rc genhtml_legend=1 00:06:33.692 --rc geninfo_all_blocks=1 00:06:33.692 --rc geninfo_unexecuted_blocks=1 00:06:33.692 00:06:33.692 ' 00:06:33.692 08:42:23 -- spdk/autotest.sh@25 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:06:33.692 08:42:23 -- nvmf/common.sh@7 -- # uname -s 00:06:33.692 08:42:23 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:33.692 08:42:23 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:33.692 08:42:23 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:33.692 08:42:23 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:33.692 08:42:23 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:33.692 08:42:23 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:33.692 08:42:23 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:33.692 08:42:23 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:33.692 08:42:23 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:33.692 08:42:23 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:33.692 08:42:23 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:06:33.693 08:42:23 -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:06:33.693 08:42:23 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:33.693 08:42:23 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:33.693 08:42:23 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:06:33.693 08:42:23 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:33.693 08:42:23 -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:06:33.693 08:42:23 -- scripts/common.sh@15 -- # shopt -s extglob 00:06:33.693 08:42:23 -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:33.693 08:42:23 -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:33.693 08:42:23 -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:33.693 08:42:23 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:33.693 08:42:23 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:33.693 08:42:23 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:33.693 08:42:23 -- paths/export.sh@5 -- # export PATH 00:06:33.693 08:42:23 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:33.693 08:42:23 -- nvmf/common.sh@51 -- # : 0 00:06:33.693 08:42:23 -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:06:33.693 08:42:23 -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:06:33.693 08:42:23 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:33.693 08:42:23 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:33.693 08:42:23 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:33.693 08:42:23 -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:06:33.693 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:06:33.693 08:42:23 -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:06:33.693 08:42:23 -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:06:33.693 08:42:23 -- nvmf/common.sh@55 -- # have_pci_nics=0 00:06:33.693 08:42:23 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:06:33.954 08:42:23 -- spdk/autotest.sh@32 -- # uname -s 00:06:33.954 08:42:23 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:06:33.954 08:42:23 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:06:33.954 08:42:23 -- spdk/autotest.sh@34 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:06:33.954 08:42:23 -- spdk/autotest.sh@39 -- # echo '|/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/core-collector.sh %P %s %t' 00:06:33.955 08:42:23 -- spdk/autotest.sh@40 -- # echo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:06:33.955 08:42:23 -- spdk/autotest.sh@44 -- # modprobe nbd 00:06:33.955 08:42:23 -- spdk/autotest.sh@46 -- # type -P udevadm 00:06:33.955 08:42:23 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:06:33.955 08:42:23 -- spdk/autotest.sh@48 -- # udevadm_pid=1632649 00:06:33.955 08:42:23 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:06:33.955 08:42:23 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:06:33.955 08:42:23 -- pm/common@17 -- # local monitor 00:06:33.955 08:42:23 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:06:33.955 08:42:23 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:06:33.955 08:42:23 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:06:33.955 08:42:23 -- pm/common@21 -- # date +%s 00:06:33.955 08:42:23 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:06:33.955 08:42:23 -- pm/common@25 -- # sleep 1 00:06:33.955 08:42:23 -- pm/common@21 -- # date +%s 00:06:33.955 08:42:23 -- pm/common@21 -- # date +%s 00:06:33.955 08:42:23 -- pm/common@21 -- # date +%s 00:06:33.955 08:42:23 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1730878943 00:06:33.955 08:42:23 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1730878943 00:06:33.955 08:42:23 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1730878943 00:06:33.955 08:42:23 -- pm/common@21 -- # sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1730878943 00:06:33.955 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1730878943_collect-cpu-load.pm.log 00:06:33.955 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1730878943_collect-vmstat.pm.log 00:06:33.955 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1730878943_collect-cpu-temp.pm.log 00:06:33.955 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1730878943_collect-bmc-pm.bmc.pm.log 00:06:34.897 08:42:24 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:06:34.897 08:42:24 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:06:34.897 08:42:24 -- common/autotest_common.sh@724 -- # xtrace_disable 00:06:34.897 08:42:24 -- common/autotest_common.sh@10 -- # set +x 00:06:34.897 08:42:24 -- spdk/autotest.sh@59 -- # create_test_list 00:06:34.897 08:42:24 -- common/autotest_common.sh@748 -- # xtrace_disable 00:06:34.897 08:42:24 -- common/autotest_common.sh@10 -- # set +x 00:06:34.897 08:42:24 -- spdk/autotest.sh@61 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/autotest.sh 00:06:34.897 08:42:24 -- spdk/autotest.sh@61 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:06:34.897 08:42:24 -- spdk/autotest.sh@61 -- # src=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:06:34.897 08:42:24 -- spdk/autotest.sh@62 -- # out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:06:34.897 08:42:24 -- spdk/autotest.sh@63 -- # cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:06:34.897 08:42:24 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:06:34.897 08:42:24 -- common/autotest_common.sh@1453 -- # uname 00:06:34.897 08:42:24 -- common/autotest_common.sh@1453 -- # '[' Linux = FreeBSD ']' 00:06:34.897 08:42:24 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:06:34.897 08:42:24 -- common/autotest_common.sh@1473 -- # uname 00:06:34.897 08:42:24 -- common/autotest_common.sh@1473 -- # [[ Linux = FreeBSD ]] 00:06:34.897 08:42:24 -- spdk/autotest.sh@68 -- # [[ y == y ]] 00:06:34.898 08:42:24 -- spdk/autotest.sh@70 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 --version 00:06:34.898 lcov: LCOV version 1.15 00:06:34.898 08:42:24 -- spdk/autotest.sh@72 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -i -t Baseline -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info 00:06:49.812 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:06:49.812 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno 00:07:07.932 08:42:55 -- spdk/autotest.sh@76 -- # timing_enter pre_cleanup 00:07:07.932 08:42:55 -- common/autotest_common.sh@724 -- # xtrace_disable 00:07:07.932 08:42:55 -- common/autotest_common.sh@10 -- # set +x 00:07:07.932 08:42:55 -- spdk/autotest.sh@78 -- # rm -f 00:07:07.932 08:42:55 -- spdk/autotest.sh@81 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:07:08.506 0000:80:01.6 (8086 0b00): Already using the ioatdma driver 00:07:08.506 0000:80:01.7 (8086 0b00): Already using the ioatdma driver 00:07:08.506 0000:80:01.4 (8086 0b00): Already using the ioatdma driver 00:07:08.506 0000:80:01.5 (8086 0b00): Already using the ioatdma driver 00:07:08.506 0000:80:01.2 (8086 0b00): Already using the ioatdma driver 00:07:08.506 0000:80:01.3 (8086 0b00): Already using the ioatdma driver 00:07:08.506 0000:80:01.0 (8086 0b00): Already using the ioatdma driver 00:07:08.506 0000:80:01.1 (8086 0b00): Already using the ioatdma driver 00:07:08.506 0000:65:00.0 (144d a80a): Already using the nvme driver 00:07:08.768 0000:00:01.6 (8086 0b00): Already using the ioatdma driver 00:07:08.768 0000:00:01.7 (8086 0b00): Already using the ioatdma driver 00:07:08.768 0000:00:01.4 (8086 0b00): Already using the ioatdma driver 00:07:08.768 0000:00:01.5 (8086 0b00): Already using the ioatdma driver 00:07:08.768 0000:00:01.2 (8086 0b00): Already using the ioatdma driver 00:07:08.768 0000:00:01.3 (8086 0b00): Already using the ioatdma driver 00:07:08.768 0000:00:01.0 (8086 0b00): Already using the ioatdma driver 00:07:08.768 0000:00:01.1 (8086 0b00): Already using the ioatdma driver 00:07:09.029 08:42:59 -- spdk/autotest.sh@83 -- # get_zoned_devs 00:07:09.029 08:42:59 -- common/autotest_common.sh@1653 -- # zoned_devs=() 00:07:09.029 08:42:59 -- common/autotest_common.sh@1653 -- # local -gA zoned_devs 00:07:09.029 08:42:59 -- common/autotest_common.sh@1654 -- # local nvme bdf 00:07:09.029 08:42:59 -- common/autotest_common.sh@1656 -- # for nvme in /sys/block/nvme* 00:07:09.029 08:42:59 -- common/autotest_common.sh@1657 -- # is_block_zoned nvme0n1 00:07:09.029 08:42:59 -- common/autotest_common.sh@1646 -- # local device=nvme0n1 00:07:09.029 08:42:59 -- common/autotest_common.sh@1648 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:07:09.029 08:42:59 -- common/autotest_common.sh@1649 -- # [[ none != none ]] 00:07:09.029 08:42:59 -- spdk/autotest.sh@85 -- # (( 0 > 0 )) 00:07:09.029 08:42:59 -- spdk/autotest.sh@97 -- # for dev in /dev/nvme*n!(*p*) 00:07:09.029 08:42:59 -- spdk/autotest.sh@99 -- # [[ -z '' ]] 00:07:09.029 08:42:59 -- spdk/autotest.sh@100 -- # block_in_use /dev/nvme0n1 00:07:09.029 08:42:59 -- scripts/common.sh@381 -- # local block=/dev/nvme0n1 pt 00:07:09.029 08:42:59 -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:07:09.029 No valid GPT data, bailing 00:07:09.291 08:42:59 -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:07:09.291 08:42:59 -- scripts/common.sh@394 -- # pt= 00:07:09.291 08:42:59 -- scripts/common.sh@395 -- # return 1 00:07:09.291 08:42:59 -- spdk/autotest.sh@101 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:07:09.291 1+0 records in 00:07:09.291 1+0 records out 00:07:09.291 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00228193 s, 460 MB/s 00:07:09.291 08:42:59 -- spdk/autotest.sh@105 -- # sync 00:07:09.291 08:42:59 -- spdk/autotest.sh@107 -- # xtrace_disable_per_cmd reap_spdk_processes 00:07:09.291 08:42:59 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:07:09.291 08:42:59 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:07:17.435 08:43:07 -- spdk/autotest.sh@111 -- # uname -s 00:07:17.435 08:43:07 -- spdk/autotest.sh@111 -- # [[ Linux == Linux ]] 00:07:17.435 08:43:07 -- spdk/autotest.sh@111 -- # [[ 0 -eq 1 ]] 00:07:17.435 08:43:07 -- spdk/autotest.sh@115 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:07:20.738 Hugepages 00:07:20.738 node hugesize free / total 00:07:20.738 node0 1048576kB 0 / 0 00:07:20.738 node0 2048kB 0 / 0 00:07:20.738 node1 1048576kB 0 / 0 00:07:20.738 node1 2048kB 0 / 0 00:07:20.738 00:07:20.738 Type BDF Vendor Device NUMA Driver Device Block devices 00:07:20.738 I/OAT 0000:00:01.0 8086 0b00 0 ioatdma - - 00:07:20.738 I/OAT 0000:00:01.1 8086 0b00 0 ioatdma - - 00:07:20.738 I/OAT 0000:00:01.2 8086 0b00 0 ioatdma - - 00:07:20.738 I/OAT 0000:00:01.3 8086 0b00 0 ioatdma - - 00:07:20.738 I/OAT 0000:00:01.4 8086 0b00 0 ioatdma - - 00:07:20.738 I/OAT 0000:00:01.5 8086 0b00 0 ioatdma - - 00:07:20.738 I/OAT 0000:00:01.6 8086 0b00 0 ioatdma - - 00:07:20.738 I/OAT 0000:00:01.7 8086 0b00 0 ioatdma - - 00:07:20.738 NVMe 0000:65:00.0 144d a80a 0 nvme nvme0 nvme0n1 00:07:20.738 I/OAT 0000:80:01.0 8086 0b00 1 ioatdma - - 00:07:20.738 I/OAT 0000:80:01.1 8086 0b00 1 ioatdma - - 00:07:20.738 I/OAT 0000:80:01.2 8086 0b00 1 ioatdma - - 00:07:20.738 I/OAT 0000:80:01.3 8086 0b00 1 ioatdma - - 00:07:20.738 I/OAT 0000:80:01.4 8086 0b00 1 ioatdma - - 00:07:20.999 I/OAT 0000:80:01.5 8086 0b00 1 ioatdma - - 00:07:20.999 I/OAT 0000:80:01.6 8086 0b00 1 ioatdma - - 00:07:20.999 I/OAT 0000:80:01.7 8086 0b00 1 ioatdma - - 00:07:20.999 08:43:10 -- spdk/autotest.sh@117 -- # uname -s 00:07:20.999 08:43:10 -- spdk/autotest.sh@117 -- # [[ Linux == Linux ]] 00:07:20.999 08:43:10 -- spdk/autotest.sh@119 -- # nvme_namespace_revert 00:07:20.999 08:43:10 -- common/autotest_common.sh@1512 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:07:24.299 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:07:24.299 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:07:26.211 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:07:26.211 08:43:16 -- common/autotest_common.sh@1513 -- # sleep 1 00:07:27.153 08:43:17 -- common/autotest_common.sh@1514 -- # bdfs=() 00:07:27.153 08:43:17 -- common/autotest_common.sh@1514 -- # local bdfs 00:07:27.153 08:43:17 -- common/autotest_common.sh@1516 -- # bdfs=($(get_nvme_bdfs)) 00:07:27.153 08:43:17 -- common/autotest_common.sh@1516 -- # get_nvme_bdfs 00:07:27.153 08:43:17 -- common/autotest_common.sh@1494 -- # bdfs=() 00:07:27.153 08:43:17 -- common/autotest_common.sh@1494 -- # local bdfs 00:07:27.153 08:43:17 -- common/autotest_common.sh@1495 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:07:27.153 08:43:17 -- common/autotest_common.sh@1495 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:07:27.153 08:43:17 -- common/autotest_common.sh@1495 -- # jq -r '.config[].params.traddr' 00:07:27.414 08:43:17 -- common/autotest_common.sh@1496 -- # (( 1 == 0 )) 00:07:27.414 08:43:17 -- common/autotest_common.sh@1500 -- # printf '%s\n' 0000:65:00.0 00:07:27.414 08:43:17 -- common/autotest_common.sh@1518 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:07:30.714 Waiting for block devices as requested 00:07:30.714 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:07:30.714 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:07:30.714 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:07:30.974 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:07:30.974 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:07:30.974 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:07:31.234 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:07:31.234 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:07:31.234 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:07:31.495 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:07:31.495 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:07:31.755 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:07:31.755 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:07:31.755 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:07:31.755 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:07:32.015 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:07:32.015 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:07:32.276 08:43:22 -- common/autotest_common.sh@1520 -- # for bdf in "${bdfs[@]}" 00:07:32.276 08:43:22 -- common/autotest_common.sh@1521 -- # get_nvme_ctrlr_from_bdf 0000:65:00.0 00:07:32.276 08:43:22 -- common/autotest_common.sh@1483 -- # readlink -f /sys/class/nvme/nvme0 00:07:32.276 08:43:22 -- common/autotest_common.sh@1483 -- # grep 0000:65:00.0/nvme/nvme 00:07:32.276 08:43:22 -- common/autotest_common.sh@1483 -- # bdf_sysfs_path=/sys/devices/pci0000:64/0000:64:02.0/0000:65:00.0/nvme/nvme0 00:07:32.276 08:43:22 -- common/autotest_common.sh@1484 -- # [[ -z /sys/devices/pci0000:64/0000:64:02.0/0000:65:00.0/nvme/nvme0 ]] 00:07:32.276 08:43:22 -- common/autotest_common.sh@1488 -- # basename /sys/devices/pci0000:64/0000:64:02.0/0000:65:00.0/nvme/nvme0 00:07:32.276 08:43:22 -- common/autotest_common.sh@1488 -- # printf '%s\n' nvme0 00:07:32.276 08:43:22 -- common/autotest_common.sh@1521 -- # nvme_ctrlr=/dev/nvme0 00:07:32.276 08:43:22 -- common/autotest_common.sh@1522 -- # [[ -z /dev/nvme0 ]] 00:07:32.276 08:43:22 -- common/autotest_common.sh@1527 -- # cut -d: -f2 00:07:32.276 08:43:22 -- common/autotest_common.sh@1527 -- # nvme id-ctrl /dev/nvme0 00:07:32.276 08:43:22 -- common/autotest_common.sh@1527 -- # grep oacs 00:07:32.276 08:43:22 -- common/autotest_common.sh@1527 -- # oacs=' 0x5f' 00:07:32.276 08:43:22 -- common/autotest_common.sh@1528 -- # oacs_ns_manage=8 00:07:32.276 08:43:22 -- common/autotest_common.sh@1530 -- # [[ 8 -ne 0 ]] 00:07:32.276 08:43:22 -- common/autotest_common.sh@1536 -- # nvme id-ctrl /dev/nvme0 00:07:32.276 08:43:22 -- common/autotest_common.sh@1536 -- # grep unvmcap 00:07:32.276 08:43:22 -- common/autotest_common.sh@1536 -- # cut -d: -f2 00:07:32.276 08:43:22 -- common/autotest_common.sh@1536 -- # unvmcap=' 0' 00:07:32.276 08:43:22 -- common/autotest_common.sh@1537 -- # [[ 0 -eq 0 ]] 00:07:32.276 08:43:22 -- common/autotest_common.sh@1539 -- # continue 00:07:32.276 08:43:22 -- spdk/autotest.sh@122 -- # timing_exit pre_cleanup 00:07:32.276 08:43:22 -- common/autotest_common.sh@730 -- # xtrace_disable 00:07:32.276 08:43:22 -- common/autotest_common.sh@10 -- # set +x 00:07:32.276 08:43:22 -- spdk/autotest.sh@125 -- # timing_enter afterboot 00:07:32.276 08:43:22 -- common/autotest_common.sh@724 -- # xtrace_disable 00:07:32.276 08:43:22 -- common/autotest_common.sh@10 -- # set +x 00:07:32.276 08:43:22 -- spdk/autotest.sh@126 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:07:35.577 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:07:35.578 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:07:35.838 08:43:25 -- spdk/autotest.sh@127 -- # timing_exit afterboot 00:07:35.838 08:43:25 -- common/autotest_common.sh@730 -- # xtrace_disable 00:07:35.838 08:43:25 -- common/autotest_common.sh@10 -- # set +x 00:07:36.098 08:43:25 -- spdk/autotest.sh@131 -- # opal_revert_cleanup 00:07:36.098 08:43:25 -- common/autotest_common.sh@1574 -- # mapfile -t bdfs 00:07:36.098 08:43:25 -- common/autotest_common.sh@1574 -- # get_nvme_bdfs_by_id 0x0a54 00:07:36.098 08:43:25 -- common/autotest_common.sh@1559 -- # bdfs=() 00:07:36.098 08:43:25 -- common/autotest_common.sh@1559 -- # _bdfs=() 00:07:36.098 08:43:25 -- common/autotest_common.sh@1559 -- # local bdfs _bdfs 00:07:36.098 08:43:25 -- common/autotest_common.sh@1560 -- # _bdfs=($(get_nvme_bdfs)) 00:07:36.098 08:43:25 -- common/autotest_common.sh@1560 -- # get_nvme_bdfs 00:07:36.098 08:43:25 -- common/autotest_common.sh@1494 -- # bdfs=() 00:07:36.098 08:43:25 -- common/autotest_common.sh@1494 -- # local bdfs 00:07:36.098 08:43:25 -- common/autotest_common.sh@1495 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:07:36.098 08:43:25 -- common/autotest_common.sh@1495 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:07:36.098 08:43:25 -- common/autotest_common.sh@1495 -- # jq -r '.config[].params.traddr' 00:07:36.098 08:43:26 -- common/autotest_common.sh@1496 -- # (( 1 == 0 )) 00:07:36.098 08:43:26 -- common/autotest_common.sh@1500 -- # printf '%s\n' 0000:65:00.0 00:07:36.098 08:43:26 -- common/autotest_common.sh@1561 -- # for bdf in "${_bdfs[@]}" 00:07:36.098 08:43:26 -- common/autotest_common.sh@1562 -- # cat /sys/bus/pci/devices/0000:65:00.0/device 00:07:36.098 08:43:26 -- common/autotest_common.sh@1562 -- # device=0xa80a 00:07:36.098 08:43:26 -- common/autotest_common.sh@1563 -- # [[ 0xa80a == \0\x\0\a\5\4 ]] 00:07:36.098 08:43:26 -- common/autotest_common.sh@1568 -- # (( 0 > 0 )) 00:07:36.098 08:43:26 -- common/autotest_common.sh@1568 -- # return 0 00:07:36.098 08:43:26 -- common/autotest_common.sh@1575 -- # [[ -z '' ]] 00:07:36.098 08:43:26 -- common/autotest_common.sh@1576 -- # return 0 00:07:36.098 08:43:26 -- spdk/autotest.sh@137 -- # '[' 0 -eq 1 ']' 00:07:36.098 08:43:26 -- spdk/autotest.sh@141 -- # '[' 1 -eq 1 ']' 00:07:36.098 08:43:26 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:07:36.098 08:43:26 -- spdk/autotest.sh@142 -- # [[ 0 -eq 1 ]] 00:07:36.098 08:43:26 -- spdk/autotest.sh@149 -- # timing_enter lib 00:07:36.098 08:43:26 -- common/autotest_common.sh@724 -- # xtrace_disable 00:07:36.098 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:07:36.098 08:43:26 -- spdk/autotest.sh@151 -- # [[ 0 -eq 1 ]] 00:07:36.098 08:43:26 -- spdk/autotest.sh@155 -- # run_test env /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:07:36.098 08:43:26 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:36.098 08:43:26 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:36.098 08:43:26 -- common/autotest_common.sh@10 -- # set +x 00:07:36.098 ************************************ 00:07:36.098 START TEST env 00:07:36.098 ************************************ 00:07:36.098 08:43:26 env -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:07:36.099 * Looking for test storage... 00:07:36.359 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env 00:07:36.359 08:43:26 env -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:07:36.359 08:43:26 env -- common/autotest_common.sh@1689 -- # lcov --version 00:07:36.359 08:43:26 env -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:07:36.359 08:43:26 env -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:07:36.359 08:43:26 env -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:36.359 08:43:26 env -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:36.359 08:43:26 env -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:36.359 08:43:26 env -- scripts/common.sh@336 -- # IFS=.-: 00:07:36.359 08:43:26 env -- scripts/common.sh@336 -- # read -ra ver1 00:07:36.359 08:43:26 env -- scripts/common.sh@337 -- # IFS=.-: 00:07:36.359 08:43:26 env -- scripts/common.sh@337 -- # read -ra ver2 00:07:36.359 08:43:26 env -- scripts/common.sh@338 -- # local 'op=<' 00:07:36.359 08:43:26 env -- scripts/common.sh@340 -- # ver1_l=2 00:07:36.359 08:43:26 env -- scripts/common.sh@341 -- # ver2_l=1 00:07:36.359 08:43:26 env -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:36.359 08:43:26 env -- scripts/common.sh@344 -- # case "$op" in 00:07:36.359 08:43:26 env -- scripts/common.sh@345 -- # : 1 00:07:36.359 08:43:26 env -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:36.359 08:43:26 env -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:36.359 08:43:26 env -- scripts/common.sh@365 -- # decimal 1 00:07:36.359 08:43:26 env -- scripts/common.sh@353 -- # local d=1 00:07:36.359 08:43:26 env -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:36.359 08:43:26 env -- scripts/common.sh@355 -- # echo 1 00:07:36.359 08:43:26 env -- scripts/common.sh@365 -- # ver1[v]=1 00:07:36.359 08:43:26 env -- scripts/common.sh@366 -- # decimal 2 00:07:36.359 08:43:26 env -- scripts/common.sh@353 -- # local d=2 00:07:36.359 08:43:26 env -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:36.359 08:43:26 env -- scripts/common.sh@355 -- # echo 2 00:07:36.359 08:43:26 env -- scripts/common.sh@366 -- # ver2[v]=2 00:07:36.359 08:43:26 env -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:36.359 08:43:26 env -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:36.359 08:43:26 env -- scripts/common.sh@368 -- # return 0 00:07:36.359 08:43:26 env -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:36.359 08:43:26 env -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:07:36.359 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:36.359 --rc genhtml_branch_coverage=1 00:07:36.359 --rc genhtml_function_coverage=1 00:07:36.359 --rc genhtml_legend=1 00:07:36.359 --rc geninfo_all_blocks=1 00:07:36.359 --rc geninfo_unexecuted_blocks=1 00:07:36.359 00:07:36.359 ' 00:07:36.359 08:43:26 env -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:07:36.359 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:36.359 --rc genhtml_branch_coverage=1 00:07:36.359 --rc genhtml_function_coverage=1 00:07:36.359 --rc genhtml_legend=1 00:07:36.359 --rc geninfo_all_blocks=1 00:07:36.359 --rc geninfo_unexecuted_blocks=1 00:07:36.359 00:07:36.360 ' 00:07:36.360 08:43:26 env -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:07:36.360 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:36.360 --rc genhtml_branch_coverage=1 00:07:36.360 --rc genhtml_function_coverage=1 00:07:36.360 --rc genhtml_legend=1 00:07:36.360 --rc geninfo_all_blocks=1 00:07:36.360 --rc geninfo_unexecuted_blocks=1 00:07:36.360 00:07:36.360 ' 00:07:36.360 08:43:26 env -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:07:36.360 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:36.360 --rc genhtml_branch_coverage=1 00:07:36.360 --rc genhtml_function_coverage=1 00:07:36.360 --rc genhtml_legend=1 00:07:36.360 --rc geninfo_all_blocks=1 00:07:36.360 --rc geninfo_unexecuted_blocks=1 00:07:36.360 00:07:36.360 ' 00:07:36.360 08:43:26 env -- env/env.sh@10 -- # run_test env_memory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:07:36.360 08:43:26 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:36.360 08:43:26 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:36.360 08:43:26 env -- common/autotest_common.sh@10 -- # set +x 00:07:36.360 ************************************ 00:07:36.360 START TEST env_memory 00:07:36.360 ************************************ 00:07:36.360 08:43:26 env.env_memory -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:07:36.360 00:07:36.360 00:07:36.360 CUnit - A unit testing framework for C - Version 2.1-3 00:07:36.360 http://cunit.sourceforge.net/ 00:07:36.360 00:07:36.360 00:07:36.360 Suite: memory 00:07:36.360 Test: alloc and free memory map ...[2024-11-06 08:43:26.405836] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:07:36.360 passed 00:07:36.360 Test: mem map translation ...[2024-11-06 08:43:26.431250] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:07:36.360 [2024-11-06 08:43:26.431270] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 595:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:07:36.360 [2024-11-06 08:43:26.431316] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 589:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:07:36.360 [2024-11-06 08:43:26.431324] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 605:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:07:36.622 passed 00:07:36.622 Test: mem map registration ...[2024-11-06 08:43:26.486436] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=200000 len=1234 00:07:36.622 [2024-11-06 08:43:26.486452] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 347:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=4d2 len=2097152 00:07:36.622 passed 00:07:36.622 Test: mem map adjacent registrations ...passed 00:07:36.622 00:07:36.622 Run Summary: Type Total Ran Passed Failed Inactive 00:07:36.622 suites 1 1 n/a 0 0 00:07:36.622 tests 4 4 4 0 0 00:07:36.622 asserts 152 152 152 0 n/a 00:07:36.622 00:07:36.622 Elapsed time = 0.192 seconds 00:07:36.622 00:07:36.622 real 0m0.206s 00:07:36.622 user 0m0.193s 00:07:36.622 sys 0m0.012s 00:07:36.622 08:43:26 env.env_memory -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:36.622 08:43:26 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:07:36.622 ************************************ 00:07:36.622 END TEST env_memory 00:07:36.622 ************************************ 00:07:36.622 08:43:26 env -- env/env.sh@11 -- # run_test env_vtophys /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:07:36.622 08:43:26 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:36.622 08:43:26 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:36.622 08:43:26 env -- common/autotest_common.sh@10 -- # set +x 00:07:36.622 ************************************ 00:07:36.622 START TEST env_vtophys 00:07:36.622 ************************************ 00:07:36.622 08:43:26 env.env_vtophys -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:07:36.622 EAL: lib.eal log level changed from notice to debug 00:07:36.622 EAL: Detected lcore 0 as core 0 on socket 0 00:07:36.622 EAL: Detected lcore 1 as core 1 on socket 0 00:07:36.622 EAL: Detected lcore 2 as core 2 on socket 0 00:07:36.622 EAL: Detected lcore 3 as core 3 on socket 0 00:07:36.622 EAL: Detected lcore 4 as core 4 on socket 0 00:07:36.622 EAL: Detected lcore 5 as core 5 on socket 0 00:07:36.622 EAL: Detected lcore 6 as core 6 on socket 0 00:07:36.622 EAL: Detected lcore 7 as core 7 on socket 0 00:07:36.622 EAL: Detected lcore 8 as core 8 on socket 0 00:07:36.622 EAL: Detected lcore 9 as core 9 on socket 0 00:07:36.622 EAL: Detected lcore 10 as core 10 on socket 0 00:07:36.622 EAL: Detected lcore 11 as core 11 on socket 0 00:07:36.622 EAL: Detected lcore 12 as core 12 on socket 0 00:07:36.622 EAL: Detected lcore 13 as core 13 on socket 0 00:07:36.622 EAL: Detected lcore 14 as core 14 on socket 0 00:07:36.622 EAL: Detected lcore 15 as core 15 on socket 0 00:07:36.622 EAL: Detected lcore 16 as core 16 on socket 0 00:07:36.622 EAL: Detected lcore 17 as core 17 on socket 0 00:07:36.622 EAL: Detected lcore 18 as core 18 on socket 0 00:07:36.622 EAL: Detected lcore 19 as core 19 on socket 0 00:07:36.622 EAL: Detected lcore 20 as core 20 on socket 0 00:07:36.622 EAL: Detected lcore 21 as core 21 on socket 0 00:07:36.622 EAL: Detected lcore 22 as core 22 on socket 0 00:07:36.622 EAL: Detected lcore 23 as core 23 on socket 0 00:07:36.622 EAL: Detected lcore 24 as core 24 on socket 0 00:07:36.622 EAL: Detected lcore 25 as core 25 on socket 0 00:07:36.622 EAL: Detected lcore 26 as core 26 on socket 0 00:07:36.622 EAL: Detected lcore 27 as core 27 on socket 0 00:07:36.622 EAL: Detected lcore 28 as core 28 on socket 0 00:07:36.622 EAL: Detected lcore 29 as core 29 on socket 0 00:07:36.622 EAL: Detected lcore 30 as core 30 on socket 0 00:07:36.622 EAL: Detected lcore 31 as core 31 on socket 0 00:07:36.622 EAL: Detected lcore 32 as core 32 on socket 0 00:07:36.622 EAL: Detected lcore 33 as core 33 on socket 0 00:07:36.622 EAL: Detected lcore 34 as core 34 on socket 0 00:07:36.622 EAL: Detected lcore 35 as core 35 on socket 0 00:07:36.622 EAL: Detected lcore 36 as core 0 on socket 1 00:07:36.622 EAL: Detected lcore 37 as core 1 on socket 1 00:07:36.622 EAL: Detected lcore 38 as core 2 on socket 1 00:07:36.622 EAL: Detected lcore 39 as core 3 on socket 1 00:07:36.622 EAL: Detected lcore 40 as core 4 on socket 1 00:07:36.622 EAL: Detected lcore 41 as core 5 on socket 1 00:07:36.622 EAL: Detected lcore 42 as core 6 on socket 1 00:07:36.622 EAL: Detected lcore 43 as core 7 on socket 1 00:07:36.622 EAL: Detected lcore 44 as core 8 on socket 1 00:07:36.622 EAL: Detected lcore 45 as core 9 on socket 1 00:07:36.622 EAL: Detected lcore 46 as core 10 on socket 1 00:07:36.622 EAL: Detected lcore 47 as core 11 on socket 1 00:07:36.622 EAL: Detected lcore 48 as core 12 on socket 1 00:07:36.622 EAL: Detected lcore 49 as core 13 on socket 1 00:07:36.622 EAL: Detected lcore 50 as core 14 on socket 1 00:07:36.622 EAL: Detected lcore 51 as core 15 on socket 1 00:07:36.622 EAL: Detected lcore 52 as core 16 on socket 1 00:07:36.622 EAL: Detected lcore 53 as core 17 on socket 1 00:07:36.622 EAL: Detected lcore 54 as core 18 on socket 1 00:07:36.622 EAL: Detected lcore 55 as core 19 on socket 1 00:07:36.622 EAL: Detected lcore 56 as core 20 on socket 1 00:07:36.622 EAL: Detected lcore 57 as core 21 on socket 1 00:07:36.622 EAL: Detected lcore 58 as core 22 on socket 1 00:07:36.622 EAL: Detected lcore 59 as core 23 on socket 1 00:07:36.622 EAL: Detected lcore 60 as core 24 on socket 1 00:07:36.622 EAL: Detected lcore 61 as core 25 on socket 1 00:07:36.622 EAL: Detected lcore 62 as core 26 on socket 1 00:07:36.622 EAL: Detected lcore 63 as core 27 on socket 1 00:07:36.622 EAL: Detected lcore 64 as core 28 on socket 1 00:07:36.622 EAL: Detected lcore 65 as core 29 on socket 1 00:07:36.622 EAL: Detected lcore 66 as core 30 on socket 1 00:07:36.622 EAL: Detected lcore 67 as core 31 on socket 1 00:07:36.622 EAL: Detected lcore 68 as core 32 on socket 1 00:07:36.622 EAL: Detected lcore 69 as core 33 on socket 1 00:07:36.622 EAL: Detected lcore 70 as core 34 on socket 1 00:07:36.622 EAL: Detected lcore 71 as core 35 on socket 1 00:07:36.622 EAL: Detected lcore 72 as core 0 on socket 0 00:07:36.623 EAL: Detected lcore 73 as core 1 on socket 0 00:07:36.623 EAL: Detected lcore 74 as core 2 on socket 0 00:07:36.623 EAL: Detected lcore 75 as core 3 on socket 0 00:07:36.623 EAL: Detected lcore 76 as core 4 on socket 0 00:07:36.623 EAL: Detected lcore 77 as core 5 on socket 0 00:07:36.623 EAL: Detected lcore 78 as core 6 on socket 0 00:07:36.623 EAL: Detected lcore 79 as core 7 on socket 0 00:07:36.623 EAL: Detected lcore 80 as core 8 on socket 0 00:07:36.623 EAL: Detected lcore 81 as core 9 on socket 0 00:07:36.623 EAL: Detected lcore 82 as core 10 on socket 0 00:07:36.623 EAL: Detected lcore 83 as core 11 on socket 0 00:07:36.623 EAL: Detected lcore 84 as core 12 on socket 0 00:07:36.623 EAL: Detected lcore 85 as core 13 on socket 0 00:07:36.623 EAL: Detected lcore 86 as core 14 on socket 0 00:07:36.623 EAL: Detected lcore 87 as core 15 on socket 0 00:07:36.623 EAL: Detected lcore 88 as core 16 on socket 0 00:07:36.623 EAL: Detected lcore 89 as core 17 on socket 0 00:07:36.623 EAL: Detected lcore 90 as core 18 on socket 0 00:07:36.623 EAL: Detected lcore 91 as core 19 on socket 0 00:07:36.623 EAL: Detected lcore 92 as core 20 on socket 0 00:07:36.623 EAL: Detected lcore 93 as core 21 on socket 0 00:07:36.623 EAL: Detected lcore 94 as core 22 on socket 0 00:07:36.623 EAL: Detected lcore 95 as core 23 on socket 0 00:07:36.623 EAL: Detected lcore 96 as core 24 on socket 0 00:07:36.623 EAL: Detected lcore 97 as core 25 on socket 0 00:07:36.623 EAL: Detected lcore 98 as core 26 on socket 0 00:07:36.623 EAL: Detected lcore 99 as core 27 on socket 0 00:07:36.623 EAL: Detected lcore 100 as core 28 on socket 0 00:07:36.623 EAL: Detected lcore 101 as core 29 on socket 0 00:07:36.623 EAL: Detected lcore 102 as core 30 on socket 0 00:07:36.623 EAL: Detected lcore 103 as core 31 on socket 0 00:07:36.623 EAL: Detected lcore 104 as core 32 on socket 0 00:07:36.623 EAL: Detected lcore 105 as core 33 on socket 0 00:07:36.623 EAL: Detected lcore 106 as core 34 on socket 0 00:07:36.623 EAL: Detected lcore 107 as core 35 on socket 0 00:07:36.623 EAL: Detected lcore 108 as core 0 on socket 1 00:07:36.623 EAL: Detected lcore 109 as core 1 on socket 1 00:07:36.623 EAL: Detected lcore 110 as core 2 on socket 1 00:07:36.623 EAL: Detected lcore 111 as core 3 on socket 1 00:07:36.623 EAL: Detected lcore 112 as core 4 on socket 1 00:07:36.623 EAL: Detected lcore 113 as core 5 on socket 1 00:07:36.623 EAL: Detected lcore 114 as core 6 on socket 1 00:07:36.623 EAL: Detected lcore 115 as core 7 on socket 1 00:07:36.623 EAL: Detected lcore 116 as core 8 on socket 1 00:07:36.623 EAL: Detected lcore 117 as core 9 on socket 1 00:07:36.623 EAL: Detected lcore 118 as core 10 on socket 1 00:07:36.623 EAL: Detected lcore 119 as core 11 on socket 1 00:07:36.623 EAL: Detected lcore 120 as core 12 on socket 1 00:07:36.623 EAL: Detected lcore 121 as core 13 on socket 1 00:07:36.623 EAL: Detected lcore 122 as core 14 on socket 1 00:07:36.623 EAL: Detected lcore 123 as core 15 on socket 1 00:07:36.623 EAL: Detected lcore 124 as core 16 on socket 1 00:07:36.623 EAL: Detected lcore 125 as core 17 on socket 1 00:07:36.623 EAL: Detected lcore 126 as core 18 on socket 1 00:07:36.623 EAL: Detected lcore 127 as core 19 on socket 1 00:07:36.623 EAL: Skipped lcore 128 as core 20 on socket 1 00:07:36.623 EAL: Skipped lcore 129 as core 21 on socket 1 00:07:36.623 EAL: Skipped lcore 130 as core 22 on socket 1 00:07:36.623 EAL: Skipped lcore 131 as core 23 on socket 1 00:07:36.623 EAL: Skipped lcore 132 as core 24 on socket 1 00:07:36.623 EAL: Skipped lcore 133 as core 25 on socket 1 00:07:36.623 EAL: Skipped lcore 134 as core 26 on socket 1 00:07:36.623 EAL: Skipped lcore 135 as core 27 on socket 1 00:07:36.623 EAL: Skipped lcore 136 as core 28 on socket 1 00:07:36.623 EAL: Skipped lcore 137 as core 29 on socket 1 00:07:36.623 EAL: Skipped lcore 138 as core 30 on socket 1 00:07:36.623 EAL: Skipped lcore 139 as core 31 on socket 1 00:07:36.623 EAL: Skipped lcore 140 as core 32 on socket 1 00:07:36.623 EAL: Skipped lcore 141 as core 33 on socket 1 00:07:36.623 EAL: Skipped lcore 142 as core 34 on socket 1 00:07:36.623 EAL: Skipped lcore 143 as core 35 on socket 1 00:07:36.623 EAL: Maximum logical cores by configuration: 128 00:07:36.623 EAL: Detected CPU lcores: 128 00:07:36.623 EAL: Detected NUMA nodes: 2 00:07:36.623 EAL: Checking presence of .so 'librte_eal.so.24.1' 00:07:36.623 EAL: Detected shared linkage of DPDK 00:07:36.623 EAL: No shared files mode enabled, IPC will be disabled 00:07:36.623 EAL: Bus pci wants IOVA as 'DC' 00:07:36.623 EAL: Buses did not request a specific IOVA mode. 00:07:36.623 EAL: IOMMU is available, selecting IOVA as VA mode. 00:07:36.623 EAL: Selected IOVA mode 'VA' 00:07:36.623 EAL: Probing VFIO support... 00:07:36.623 EAL: IOMMU type 1 (Type 1) is supported 00:07:36.623 EAL: IOMMU type 7 (sPAPR) is not supported 00:07:36.623 EAL: IOMMU type 8 (No-IOMMU) is not supported 00:07:36.623 EAL: VFIO support initialized 00:07:36.623 EAL: Ask a virtual area of 0x2e000 bytes 00:07:36.623 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:07:36.623 EAL: Setting up physically contiguous memory... 00:07:36.623 EAL: Setting maximum number of open files to 524288 00:07:36.623 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:07:36.623 EAL: Detected memory type: socket_id:1 hugepage_sz:2097152 00:07:36.623 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:07:36.623 EAL: Ask a virtual area of 0x61000 bytes 00:07:36.623 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:07:36.623 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:07:36.623 EAL: Ask a virtual area of 0x400000000 bytes 00:07:36.623 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:07:36.623 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:07:36.623 EAL: Ask a virtual area of 0x61000 bytes 00:07:36.623 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:07:36.623 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:07:36.623 EAL: Ask a virtual area of 0x400000000 bytes 00:07:36.623 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:07:36.623 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:07:36.623 EAL: Ask a virtual area of 0x61000 bytes 00:07:36.623 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:07:36.623 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:07:36.623 EAL: Ask a virtual area of 0x400000000 bytes 00:07:36.623 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:07:36.623 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:07:36.623 EAL: Ask a virtual area of 0x61000 bytes 00:07:36.623 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:07:36.623 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:07:36.623 EAL: Ask a virtual area of 0x400000000 bytes 00:07:36.623 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:07:36.623 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:07:36.623 EAL: Creating 4 segment lists: n_segs:8192 socket_id:1 hugepage_sz:2097152 00:07:36.623 EAL: Ask a virtual area of 0x61000 bytes 00:07:36.623 EAL: Virtual area found at 0x201000800000 (size = 0x61000) 00:07:36.623 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:07:36.623 EAL: Ask a virtual area of 0x400000000 bytes 00:07:36.623 EAL: Virtual area found at 0x201000a00000 (size = 0x400000000) 00:07:36.623 EAL: VA reserved for memseg list at 0x201000a00000, size 400000000 00:07:36.623 EAL: Ask a virtual area of 0x61000 bytes 00:07:36.623 EAL: Virtual area found at 0x201400a00000 (size = 0x61000) 00:07:36.623 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:07:36.623 EAL: Ask a virtual area of 0x400000000 bytes 00:07:36.623 EAL: Virtual area found at 0x201400c00000 (size = 0x400000000) 00:07:36.623 EAL: VA reserved for memseg list at 0x201400c00000, size 400000000 00:07:36.623 EAL: Ask a virtual area of 0x61000 bytes 00:07:36.623 EAL: Virtual area found at 0x201800c00000 (size = 0x61000) 00:07:36.623 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:07:36.623 EAL: Ask a virtual area of 0x400000000 bytes 00:07:36.623 EAL: Virtual area found at 0x201800e00000 (size = 0x400000000) 00:07:36.623 EAL: VA reserved for memseg list at 0x201800e00000, size 400000000 00:07:36.623 EAL: Ask a virtual area of 0x61000 bytes 00:07:36.623 EAL: Virtual area found at 0x201c00e00000 (size = 0x61000) 00:07:36.623 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:07:36.623 EAL: Ask a virtual area of 0x400000000 bytes 00:07:36.623 EAL: Virtual area found at 0x201c01000000 (size = 0x400000000) 00:07:36.623 EAL: VA reserved for memseg list at 0x201c01000000, size 400000000 00:07:36.623 EAL: Hugepages will be freed exactly as allocated. 00:07:36.623 EAL: No shared files mode enabled, IPC is disabled 00:07:36.623 EAL: No shared files mode enabled, IPC is disabled 00:07:36.623 EAL: TSC frequency is ~2400000 KHz 00:07:36.623 EAL: Main lcore 0 is ready (tid=7f4966471a00;cpuset=[0]) 00:07:36.623 EAL: Trying to obtain current memory policy. 00:07:36.623 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:36.623 EAL: Restoring previous memory policy: 0 00:07:36.623 EAL: request: mp_malloc_sync 00:07:36.623 EAL: No shared files mode enabled, IPC is disabled 00:07:36.623 EAL: Heap on socket 0 was expanded by 2MB 00:07:36.623 EAL: No shared files mode enabled, IPC is disabled 00:07:36.623 EAL: No PCI address specified using 'addr=' in: bus=pci 00:07:36.623 EAL: Mem event callback 'spdk:(nil)' registered 00:07:36.885 00:07:36.885 00:07:36.885 CUnit - A unit testing framework for C - Version 2.1-3 00:07:36.885 http://cunit.sourceforge.net/ 00:07:36.885 00:07:36.885 00:07:36.885 Suite: components_suite 00:07:36.885 Test: vtophys_malloc_test ...passed 00:07:36.885 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:07:36.885 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:36.885 EAL: Restoring previous memory policy: 4 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was expanded by 4MB 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was shrunk by 4MB 00:07:36.885 EAL: Trying to obtain current memory policy. 00:07:36.885 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:36.885 EAL: Restoring previous memory policy: 4 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was expanded by 6MB 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was shrunk by 6MB 00:07:36.885 EAL: Trying to obtain current memory policy. 00:07:36.885 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:36.885 EAL: Restoring previous memory policy: 4 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was expanded by 10MB 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was shrunk by 10MB 00:07:36.885 EAL: Trying to obtain current memory policy. 00:07:36.885 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:36.885 EAL: Restoring previous memory policy: 4 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was expanded by 18MB 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was shrunk by 18MB 00:07:36.885 EAL: Trying to obtain current memory policy. 00:07:36.885 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:36.885 EAL: Restoring previous memory policy: 4 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was expanded by 34MB 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was shrunk by 34MB 00:07:36.885 EAL: Trying to obtain current memory policy. 00:07:36.885 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:36.885 EAL: Restoring previous memory policy: 4 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was expanded by 66MB 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was shrunk by 66MB 00:07:36.885 EAL: Trying to obtain current memory policy. 00:07:36.885 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:36.885 EAL: Restoring previous memory policy: 4 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was expanded by 130MB 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was shrunk by 130MB 00:07:36.885 EAL: Trying to obtain current memory policy. 00:07:36.885 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:36.885 EAL: Restoring previous memory policy: 4 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was expanded by 258MB 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was shrunk by 258MB 00:07:36.885 EAL: Trying to obtain current memory policy. 00:07:36.885 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:36.885 EAL: Restoring previous memory policy: 4 00:07:36.885 EAL: Calling mem event callback 'spdk:(nil)' 00:07:36.885 EAL: request: mp_malloc_sync 00:07:36.885 EAL: No shared files mode enabled, IPC is disabled 00:07:36.885 EAL: Heap on socket 0 was expanded by 514MB 00:07:37.146 EAL: Calling mem event callback 'spdk:(nil)' 00:07:37.146 EAL: request: mp_malloc_sync 00:07:37.146 EAL: No shared files mode enabled, IPC is disabled 00:07:37.146 EAL: Heap on socket 0 was shrunk by 514MB 00:07:37.146 EAL: Trying to obtain current memory policy. 00:07:37.146 EAL: Setting policy MPOL_PREFERRED for socket 0 00:07:37.146 EAL: Restoring previous memory policy: 4 00:07:37.146 EAL: Calling mem event callback 'spdk:(nil)' 00:07:37.146 EAL: request: mp_malloc_sync 00:07:37.146 EAL: No shared files mode enabled, IPC is disabled 00:07:37.146 EAL: Heap on socket 0 was expanded by 1026MB 00:07:37.406 EAL: Calling mem event callback 'spdk:(nil)' 00:07:37.406 EAL: request: mp_malloc_sync 00:07:37.406 EAL: No shared files mode enabled, IPC is disabled 00:07:37.406 EAL: Heap on socket 0 was shrunk by 1026MB 00:07:37.406 passed 00:07:37.406 00:07:37.406 Run Summary: Type Total Ran Passed Failed Inactive 00:07:37.406 suites 1 1 n/a 0 0 00:07:37.406 tests 2 2 2 0 0 00:07:37.406 asserts 497 497 497 0 n/a 00:07:37.406 00:07:37.406 Elapsed time = 0.657 seconds 00:07:37.406 EAL: Calling mem event callback 'spdk:(nil)' 00:07:37.406 EAL: request: mp_malloc_sync 00:07:37.406 EAL: No shared files mode enabled, IPC is disabled 00:07:37.406 EAL: Heap on socket 0 was shrunk by 2MB 00:07:37.406 EAL: No shared files mode enabled, IPC is disabled 00:07:37.406 EAL: No shared files mode enabled, IPC is disabled 00:07:37.406 EAL: No shared files mode enabled, IPC is disabled 00:07:37.406 00:07:37.406 real 0m0.794s 00:07:37.406 user 0m0.421s 00:07:37.406 sys 0m0.343s 00:07:37.406 08:43:27 env.env_vtophys -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:37.407 08:43:27 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:07:37.407 ************************************ 00:07:37.407 END TEST env_vtophys 00:07:37.407 ************************************ 00:07:37.407 08:43:27 env -- env/env.sh@12 -- # run_test env_pci /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:07:37.407 08:43:27 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:37.407 08:43:27 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:37.407 08:43:27 env -- common/autotest_common.sh@10 -- # set +x 00:07:37.407 ************************************ 00:07:37.407 START TEST env_pci 00:07:37.407 ************************************ 00:07:37.407 08:43:27 env.env_pci -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:07:37.667 00:07:37.667 00:07:37.667 CUnit - A unit testing framework for C - Version 2.1-3 00:07:37.667 http://cunit.sourceforge.net/ 00:07:37.667 00:07:37.667 00:07:37.667 Suite: pci 00:07:37.668 Test: pci_hook ...[2024-11-06 08:43:27.534079] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/pci.c:1049:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 1651724 has claimed it 00:07:37.668 EAL: Cannot find device (10000:00:01.0) 00:07:37.668 EAL: Failed to attach device on primary process 00:07:37.668 passed 00:07:37.668 00:07:37.668 Run Summary: Type Total Ran Passed Failed Inactive 00:07:37.668 suites 1 1 n/a 0 0 00:07:37.668 tests 1 1 1 0 0 00:07:37.668 asserts 25 25 25 0 n/a 00:07:37.668 00:07:37.668 Elapsed time = 0.029 seconds 00:07:37.668 00:07:37.668 real 0m0.050s 00:07:37.668 user 0m0.014s 00:07:37.668 sys 0m0.035s 00:07:37.668 08:43:27 env.env_pci -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:37.668 08:43:27 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:07:37.668 ************************************ 00:07:37.668 END TEST env_pci 00:07:37.668 ************************************ 00:07:37.668 08:43:27 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:07:37.668 08:43:27 env -- env/env.sh@15 -- # uname 00:07:37.668 08:43:27 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:07:37.668 08:43:27 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:07:37.668 08:43:27 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:07:37.668 08:43:27 env -- common/autotest_common.sh@1101 -- # '[' 5 -le 1 ']' 00:07:37.668 08:43:27 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:37.668 08:43:27 env -- common/autotest_common.sh@10 -- # set +x 00:07:37.668 ************************************ 00:07:37.668 START TEST env_dpdk_post_init 00:07:37.668 ************************************ 00:07:37.668 08:43:27 env.env_dpdk_post_init -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:07:37.668 EAL: Detected CPU lcores: 128 00:07:37.668 EAL: Detected NUMA nodes: 2 00:07:37.668 EAL: Detected shared linkage of DPDK 00:07:37.668 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:07:37.668 EAL: Selected IOVA mode 'VA' 00:07:37.668 EAL: VFIO support initialized 00:07:37.668 TELEMETRY: No legacy callbacks, legacy socket not created 00:07:37.668 EAL: Using IOMMU type 1 (Type 1) 00:07:37.928 EAL: Ignore mapping IO port bar(1) 00:07:37.928 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.0 (socket 0) 00:07:38.189 EAL: Ignore mapping IO port bar(1) 00:07:38.189 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.1 (socket 0) 00:07:38.450 EAL: Ignore mapping IO port bar(1) 00:07:38.450 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.2 (socket 0) 00:07:38.710 EAL: Ignore mapping IO port bar(1) 00:07:38.710 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.3 (socket 0) 00:07:38.710 EAL: Ignore mapping IO port bar(1) 00:07:38.970 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.4 (socket 0) 00:07:38.970 EAL: Ignore mapping IO port bar(1) 00:07:39.231 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.5 (socket 0) 00:07:39.231 EAL: Ignore mapping IO port bar(1) 00:07:39.231 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.6 (socket 0) 00:07:39.491 EAL: Ignore mapping IO port bar(1) 00:07:39.491 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:00:01.7 (socket 0) 00:07:39.750 EAL: Probe PCI driver: spdk_nvme (144d:a80a) device: 0000:65:00.0 (socket 0) 00:07:40.011 EAL: Ignore mapping IO port bar(1) 00:07:40.011 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.0 (socket 1) 00:07:40.271 EAL: Ignore mapping IO port bar(1) 00:07:40.271 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.1 (socket 1) 00:07:40.271 EAL: Ignore mapping IO port bar(1) 00:07:40.530 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.2 (socket 1) 00:07:40.530 EAL: Ignore mapping IO port bar(1) 00:07:40.790 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.3 (socket 1) 00:07:40.791 EAL: Ignore mapping IO port bar(1) 00:07:41.051 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.4 (socket 1) 00:07:41.051 EAL: Ignore mapping IO port bar(1) 00:07:41.051 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.5 (socket 1) 00:07:41.312 EAL: Ignore mapping IO port bar(1) 00:07:41.312 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.6 (socket 1) 00:07:41.572 EAL: Ignore mapping IO port bar(1) 00:07:41.572 EAL: Probe PCI driver: spdk_ioat (8086:0b00) device: 0000:80:01.7 (socket 1) 00:07:41.572 EAL: Releasing PCI mapped resource for 0000:65:00.0 00:07:41.572 EAL: Calling pci_unmap_resource for 0000:65:00.0 at 0x202001020000 00:07:41.832 Starting DPDK initialization... 00:07:41.832 Starting SPDK post initialization... 00:07:41.832 SPDK NVMe probe 00:07:41.832 Attaching to 0000:65:00.0 00:07:41.832 Attached to 0000:65:00.0 00:07:41.832 Cleaning up... 00:07:43.744 00:07:43.744 real 0m5.720s 00:07:43.744 user 0m0.103s 00:07:43.744 sys 0m0.168s 00:07:43.744 08:43:33 env.env_dpdk_post_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:43.744 08:43:33 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:07:43.744 ************************************ 00:07:43.744 END TEST env_dpdk_post_init 00:07:43.744 ************************************ 00:07:43.744 08:43:33 env -- env/env.sh@26 -- # uname 00:07:43.744 08:43:33 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:07:43.744 08:43:33 env -- env/env.sh@29 -- # run_test env_mem_callbacks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:07:43.744 08:43:33 env -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:43.744 08:43:33 env -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:43.744 08:43:33 env -- common/autotest_common.sh@10 -- # set +x 00:07:43.744 ************************************ 00:07:43.744 START TEST env_mem_callbacks 00:07:43.745 ************************************ 00:07:43.745 08:43:33 env.env_mem_callbacks -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:07:43.745 EAL: Detected CPU lcores: 128 00:07:43.745 EAL: Detected NUMA nodes: 2 00:07:43.745 EAL: Detected shared linkage of DPDK 00:07:43.745 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:07:43.745 EAL: Selected IOVA mode 'VA' 00:07:43.745 EAL: VFIO support initialized 00:07:43.745 TELEMETRY: No legacy callbacks, legacy socket not created 00:07:43.745 00:07:43.745 00:07:43.745 CUnit - A unit testing framework for C - Version 2.1-3 00:07:43.745 http://cunit.sourceforge.net/ 00:07:43.745 00:07:43.745 00:07:43.745 Suite: memory 00:07:43.745 Test: test ... 00:07:43.745 register 0x200000200000 2097152 00:07:43.745 malloc 3145728 00:07:43.745 register 0x200000400000 4194304 00:07:43.745 buf 0x200000500000 len 3145728 PASSED 00:07:43.745 malloc 64 00:07:43.745 buf 0x2000004fff40 len 64 PASSED 00:07:43.745 malloc 4194304 00:07:43.745 register 0x200000800000 6291456 00:07:43.745 buf 0x200000a00000 len 4194304 PASSED 00:07:43.745 free 0x200000500000 3145728 00:07:43.745 free 0x2000004fff40 64 00:07:43.745 unregister 0x200000400000 4194304 PASSED 00:07:43.745 free 0x200000a00000 4194304 00:07:43.745 unregister 0x200000800000 6291456 PASSED 00:07:43.745 malloc 8388608 00:07:43.745 register 0x200000400000 10485760 00:07:43.745 buf 0x200000600000 len 8388608 PASSED 00:07:43.745 free 0x200000600000 8388608 00:07:43.745 unregister 0x200000400000 10485760 PASSED 00:07:43.745 passed 00:07:43.745 00:07:43.745 Run Summary: Type Total Ran Passed Failed Inactive 00:07:43.745 suites 1 1 n/a 0 0 00:07:43.745 tests 1 1 1 0 0 00:07:43.745 asserts 15 15 15 0 n/a 00:07:43.745 00:07:43.745 Elapsed time = 0.004 seconds 00:07:43.745 00:07:43.745 real 0m0.059s 00:07:43.745 user 0m0.017s 00:07:43.745 sys 0m0.042s 00:07:43.745 08:43:33 env.env_mem_callbacks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:43.745 08:43:33 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:07:43.745 ************************************ 00:07:43.745 END TEST env_mem_callbacks 00:07:43.745 ************************************ 00:07:43.745 00:07:43.745 real 0m7.446s 00:07:43.745 user 0m1.024s 00:07:43.745 sys 0m0.976s 00:07:43.745 08:43:33 env -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:43.745 08:43:33 env -- common/autotest_common.sh@10 -- # set +x 00:07:43.745 ************************************ 00:07:43.745 END TEST env 00:07:43.745 ************************************ 00:07:43.745 08:43:33 -- spdk/autotest.sh@156 -- # run_test rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:07:43.745 08:43:33 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:43.745 08:43:33 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:43.745 08:43:33 -- common/autotest_common.sh@10 -- # set +x 00:07:43.745 ************************************ 00:07:43.745 START TEST rpc 00:07:43.745 ************************************ 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:07:43.745 * Looking for test storage... 00:07:43.745 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@1689 -- # lcov --version 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:07:43.745 08:43:33 rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:43.745 08:43:33 rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:43.745 08:43:33 rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:43.745 08:43:33 rpc -- scripts/common.sh@336 -- # IFS=.-: 00:07:43.745 08:43:33 rpc -- scripts/common.sh@336 -- # read -ra ver1 00:07:43.745 08:43:33 rpc -- scripts/common.sh@337 -- # IFS=.-: 00:07:43.745 08:43:33 rpc -- scripts/common.sh@337 -- # read -ra ver2 00:07:43.745 08:43:33 rpc -- scripts/common.sh@338 -- # local 'op=<' 00:07:43.745 08:43:33 rpc -- scripts/common.sh@340 -- # ver1_l=2 00:07:43.745 08:43:33 rpc -- scripts/common.sh@341 -- # ver2_l=1 00:07:43.745 08:43:33 rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:43.745 08:43:33 rpc -- scripts/common.sh@344 -- # case "$op" in 00:07:43.745 08:43:33 rpc -- scripts/common.sh@345 -- # : 1 00:07:43.745 08:43:33 rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:43.745 08:43:33 rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:43.745 08:43:33 rpc -- scripts/common.sh@365 -- # decimal 1 00:07:43.745 08:43:33 rpc -- scripts/common.sh@353 -- # local d=1 00:07:43.745 08:43:33 rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:43.745 08:43:33 rpc -- scripts/common.sh@355 -- # echo 1 00:07:43.745 08:43:33 rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:07:43.745 08:43:33 rpc -- scripts/common.sh@366 -- # decimal 2 00:07:43.745 08:43:33 rpc -- scripts/common.sh@353 -- # local d=2 00:07:43.745 08:43:33 rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:43.745 08:43:33 rpc -- scripts/common.sh@355 -- # echo 2 00:07:43.745 08:43:33 rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:07:43.745 08:43:33 rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:43.745 08:43:33 rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:43.745 08:43:33 rpc -- scripts/common.sh@368 -- # return 0 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:07:43.745 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:43.745 --rc genhtml_branch_coverage=1 00:07:43.745 --rc genhtml_function_coverage=1 00:07:43.745 --rc genhtml_legend=1 00:07:43.745 --rc geninfo_all_blocks=1 00:07:43.745 --rc geninfo_unexecuted_blocks=1 00:07:43.745 00:07:43.745 ' 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:07:43.745 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:43.745 --rc genhtml_branch_coverage=1 00:07:43.745 --rc genhtml_function_coverage=1 00:07:43.745 --rc genhtml_legend=1 00:07:43.745 --rc geninfo_all_blocks=1 00:07:43.745 --rc geninfo_unexecuted_blocks=1 00:07:43.745 00:07:43.745 ' 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:07:43.745 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:43.745 --rc genhtml_branch_coverage=1 00:07:43.745 --rc genhtml_function_coverage=1 00:07:43.745 --rc genhtml_legend=1 00:07:43.745 --rc geninfo_all_blocks=1 00:07:43.745 --rc geninfo_unexecuted_blocks=1 00:07:43.745 00:07:43.745 ' 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:07:43.745 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:43.745 --rc genhtml_branch_coverage=1 00:07:43.745 --rc genhtml_function_coverage=1 00:07:43.745 --rc genhtml_legend=1 00:07:43.745 --rc geninfo_all_blocks=1 00:07:43.745 --rc geninfo_unexecuted_blocks=1 00:07:43.745 00:07:43.745 ' 00:07:43.745 08:43:33 rpc -- rpc/rpc.sh@65 -- # spdk_pid=1653176 00:07:43.745 08:43:33 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:07:43.745 08:43:33 rpc -- rpc/rpc.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -e bdev 00:07:43.745 08:43:33 rpc -- rpc/rpc.sh@67 -- # waitforlisten 1653176 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@831 -- # '[' -z 1653176 ']' 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:43.745 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:43.745 08:43:33 rpc -- common/autotest_common.sh@10 -- # set +x 00:07:44.006 [2024-11-06 08:43:33.895479] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:07:44.006 [2024-11-06 08:43:33.895547] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1653176 ] 00:07:44.006 [2024-11-06 08:43:33.968111] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:44.006 [2024-11-06 08:43:34.003172] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:07:44.006 [2024-11-06 08:43:34.003207] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 1653176' to capture a snapshot of events at runtime. 00:07:44.006 [2024-11-06 08:43:34.003215] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:44.006 [2024-11-06 08:43:34.003222] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:44.006 [2024-11-06 08:43:34.003228] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid1653176 for offline analysis/debug. 00:07:44.006 [2024-11-06 08:43:34.003801] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:44.577 08:43:34 rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:44.577 08:43:34 rpc -- common/autotest_common.sh@864 -- # return 0 00:07:44.577 08:43:34 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:07:44.577 08:43:34 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:07:44.577 08:43:34 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:07:44.577 08:43:34 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:07:44.577 08:43:34 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:44.577 08:43:34 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:44.577 08:43:34 rpc -- common/autotest_common.sh@10 -- # set +x 00:07:44.838 ************************************ 00:07:44.838 START TEST rpc_integrity 00:07:44.838 ************************************ 00:07:44.838 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:07:44.838 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:07:44.838 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.838 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:44.838 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.838 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:07:44.838 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:07:44.838 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:07:44.838 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:07:44.838 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.838 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:44.838 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.838 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:07:44.838 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:07:44.838 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.838 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:44.838 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.838 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:07:44.838 { 00:07:44.838 "name": "Malloc0", 00:07:44.838 "aliases": [ 00:07:44.838 "ec6d81c4-74c9-4093-816b-9ef6bfe81ce3" 00:07:44.838 ], 00:07:44.838 "product_name": "Malloc disk", 00:07:44.838 "block_size": 512, 00:07:44.838 "num_blocks": 16384, 00:07:44.838 "uuid": "ec6d81c4-74c9-4093-816b-9ef6bfe81ce3", 00:07:44.838 "assigned_rate_limits": { 00:07:44.838 "rw_ios_per_sec": 0, 00:07:44.838 "rw_mbytes_per_sec": 0, 00:07:44.838 "r_mbytes_per_sec": 0, 00:07:44.838 "w_mbytes_per_sec": 0 00:07:44.838 }, 00:07:44.838 "claimed": false, 00:07:44.838 "zoned": false, 00:07:44.838 "supported_io_types": { 00:07:44.838 "read": true, 00:07:44.838 "write": true, 00:07:44.838 "unmap": true, 00:07:44.838 "flush": true, 00:07:44.838 "reset": true, 00:07:44.838 "nvme_admin": false, 00:07:44.838 "nvme_io": false, 00:07:44.838 "nvme_io_md": false, 00:07:44.838 "write_zeroes": true, 00:07:44.838 "zcopy": true, 00:07:44.838 "get_zone_info": false, 00:07:44.838 "zone_management": false, 00:07:44.838 "zone_append": false, 00:07:44.838 "compare": false, 00:07:44.838 "compare_and_write": false, 00:07:44.838 "abort": true, 00:07:44.838 "seek_hole": false, 00:07:44.838 "seek_data": false, 00:07:44.838 "copy": true, 00:07:44.838 "nvme_iov_md": false 00:07:44.838 }, 00:07:44.838 "memory_domains": [ 00:07:44.838 { 00:07:44.838 "dma_device_id": "system", 00:07:44.838 "dma_device_type": 1 00:07:44.838 }, 00:07:44.838 { 00:07:44.838 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:44.838 "dma_device_type": 2 00:07:44.838 } 00:07:44.838 ], 00:07:44.838 "driver_specific": {} 00:07:44.838 } 00:07:44.838 ]' 00:07:44.838 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:07:44.838 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:07:44.839 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:44.839 [2024-11-06 08:43:34.860011] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:07:44.839 [2024-11-06 08:43:34.860044] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:44.839 [2024-11-06 08:43:34.860058] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xd1d800 00:07:44.839 [2024-11-06 08:43:34.860065] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:44.839 [2024-11-06 08:43:34.861413] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:44.839 [2024-11-06 08:43:34.861435] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:07:44.839 Passthru0 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.839 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.839 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:07:44.839 { 00:07:44.839 "name": "Malloc0", 00:07:44.839 "aliases": [ 00:07:44.839 "ec6d81c4-74c9-4093-816b-9ef6bfe81ce3" 00:07:44.839 ], 00:07:44.839 "product_name": "Malloc disk", 00:07:44.839 "block_size": 512, 00:07:44.839 "num_blocks": 16384, 00:07:44.839 "uuid": "ec6d81c4-74c9-4093-816b-9ef6bfe81ce3", 00:07:44.839 "assigned_rate_limits": { 00:07:44.839 "rw_ios_per_sec": 0, 00:07:44.839 "rw_mbytes_per_sec": 0, 00:07:44.839 "r_mbytes_per_sec": 0, 00:07:44.839 "w_mbytes_per_sec": 0 00:07:44.839 }, 00:07:44.839 "claimed": true, 00:07:44.839 "claim_type": "exclusive_write", 00:07:44.839 "zoned": false, 00:07:44.839 "supported_io_types": { 00:07:44.839 "read": true, 00:07:44.839 "write": true, 00:07:44.839 "unmap": true, 00:07:44.839 "flush": true, 00:07:44.839 "reset": true, 00:07:44.839 "nvme_admin": false, 00:07:44.839 "nvme_io": false, 00:07:44.839 "nvme_io_md": false, 00:07:44.839 "write_zeroes": true, 00:07:44.839 "zcopy": true, 00:07:44.839 "get_zone_info": false, 00:07:44.839 "zone_management": false, 00:07:44.839 "zone_append": false, 00:07:44.839 "compare": false, 00:07:44.839 "compare_and_write": false, 00:07:44.839 "abort": true, 00:07:44.839 "seek_hole": false, 00:07:44.839 "seek_data": false, 00:07:44.839 "copy": true, 00:07:44.839 "nvme_iov_md": false 00:07:44.839 }, 00:07:44.839 "memory_domains": [ 00:07:44.839 { 00:07:44.839 "dma_device_id": "system", 00:07:44.839 "dma_device_type": 1 00:07:44.839 }, 00:07:44.839 { 00:07:44.839 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:44.839 "dma_device_type": 2 00:07:44.839 } 00:07:44.839 ], 00:07:44.839 "driver_specific": {} 00:07:44.839 }, 00:07:44.839 { 00:07:44.839 "name": "Passthru0", 00:07:44.839 "aliases": [ 00:07:44.839 "87eff1ef-1253-575d-b48a-239b682d3173" 00:07:44.839 ], 00:07:44.839 "product_name": "passthru", 00:07:44.839 "block_size": 512, 00:07:44.839 "num_blocks": 16384, 00:07:44.839 "uuid": "87eff1ef-1253-575d-b48a-239b682d3173", 00:07:44.839 "assigned_rate_limits": { 00:07:44.839 "rw_ios_per_sec": 0, 00:07:44.839 "rw_mbytes_per_sec": 0, 00:07:44.839 "r_mbytes_per_sec": 0, 00:07:44.839 "w_mbytes_per_sec": 0 00:07:44.839 }, 00:07:44.839 "claimed": false, 00:07:44.839 "zoned": false, 00:07:44.839 "supported_io_types": { 00:07:44.839 "read": true, 00:07:44.839 "write": true, 00:07:44.839 "unmap": true, 00:07:44.839 "flush": true, 00:07:44.839 "reset": true, 00:07:44.839 "nvme_admin": false, 00:07:44.839 "nvme_io": false, 00:07:44.839 "nvme_io_md": false, 00:07:44.839 "write_zeroes": true, 00:07:44.839 "zcopy": true, 00:07:44.839 "get_zone_info": false, 00:07:44.839 "zone_management": false, 00:07:44.839 "zone_append": false, 00:07:44.839 "compare": false, 00:07:44.839 "compare_and_write": false, 00:07:44.839 "abort": true, 00:07:44.839 "seek_hole": false, 00:07:44.839 "seek_data": false, 00:07:44.839 "copy": true, 00:07:44.839 "nvme_iov_md": false 00:07:44.839 }, 00:07:44.839 "memory_domains": [ 00:07:44.839 { 00:07:44.839 "dma_device_id": "system", 00:07:44.839 "dma_device_type": 1 00:07:44.839 }, 00:07:44.839 { 00:07:44.839 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:44.839 "dma_device_type": 2 00:07:44.839 } 00:07:44.839 ], 00:07:44.839 "driver_specific": { 00:07:44.839 "passthru": { 00:07:44.839 "name": "Passthru0", 00:07:44.839 "base_bdev_name": "Malloc0" 00:07:44.839 } 00:07:44.839 } 00:07:44.839 } 00:07:44.839 ]' 00:07:44.839 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:07:44.839 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:07:44.839 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:44.839 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:44.839 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.100 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.100 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:07:45.100 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.100 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.100 08:43:34 rpc.rpc_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.100 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:07:45.100 08:43:34 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:07:45.100 08:43:35 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:07:45.100 00:07:45.100 real 0m0.292s 00:07:45.100 user 0m0.179s 00:07:45.100 sys 0m0.043s 00:07:45.100 08:43:35 rpc.rpc_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:45.100 08:43:35 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.100 ************************************ 00:07:45.100 END TEST rpc_integrity 00:07:45.100 ************************************ 00:07:45.100 08:43:35 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:07:45.100 08:43:35 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:45.100 08:43:35 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:45.100 08:43:35 rpc -- common/autotest_common.sh@10 -- # set +x 00:07:45.100 ************************************ 00:07:45.100 START TEST rpc_plugins 00:07:45.100 ************************************ 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@1125 -- # rpc_plugins 00:07:45.100 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.100 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:07:45.100 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.100 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:07:45.100 { 00:07:45.100 "name": "Malloc1", 00:07:45.100 "aliases": [ 00:07:45.100 "7955298c-1e9e-41b8-bbc9-d6ea101f9b72" 00:07:45.100 ], 00:07:45.100 "product_name": "Malloc disk", 00:07:45.100 "block_size": 4096, 00:07:45.100 "num_blocks": 256, 00:07:45.100 "uuid": "7955298c-1e9e-41b8-bbc9-d6ea101f9b72", 00:07:45.100 "assigned_rate_limits": { 00:07:45.100 "rw_ios_per_sec": 0, 00:07:45.100 "rw_mbytes_per_sec": 0, 00:07:45.100 "r_mbytes_per_sec": 0, 00:07:45.100 "w_mbytes_per_sec": 0 00:07:45.100 }, 00:07:45.100 "claimed": false, 00:07:45.100 "zoned": false, 00:07:45.100 "supported_io_types": { 00:07:45.100 "read": true, 00:07:45.100 "write": true, 00:07:45.100 "unmap": true, 00:07:45.100 "flush": true, 00:07:45.100 "reset": true, 00:07:45.100 "nvme_admin": false, 00:07:45.100 "nvme_io": false, 00:07:45.100 "nvme_io_md": false, 00:07:45.100 "write_zeroes": true, 00:07:45.100 "zcopy": true, 00:07:45.100 "get_zone_info": false, 00:07:45.100 "zone_management": false, 00:07:45.100 "zone_append": false, 00:07:45.100 "compare": false, 00:07:45.100 "compare_and_write": false, 00:07:45.100 "abort": true, 00:07:45.100 "seek_hole": false, 00:07:45.100 "seek_data": false, 00:07:45.100 "copy": true, 00:07:45.100 "nvme_iov_md": false 00:07:45.100 }, 00:07:45.100 "memory_domains": [ 00:07:45.100 { 00:07:45.100 "dma_device_id": "system", 00:07:45.100 "dma_device_type": 1 00:07:45.100 }, 00:07:45.100 { 00:07:45.100 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:45.100 "dma_device_type": 2 00:07:45.100 } 00:07:45.100 ], 00:07:45.100 "driver_specific": {} 00:07:45.100 } 00:07:45.100 ]' 00:07:45.100 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:07:45.100 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:07:45.100 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.100 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:07:45.100 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.100 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:07:45.100 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:07:45.360 08:43:35 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:07:45.360 00:07:45.360 real 0m0.154s 00:07:45.360 user 0m0.094s 00:07:45.360 sys 0m0.021s 00:07:45.360 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:45.360 08:43:35 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:07:45.360 ************************************ 00:07:45.360 END TEST rpc_plugins 00:07:45.360 ************************************ 00:07:45.360 08:43:35 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:07:45.360 08:43:35 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:45.360 08:43:35 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:45.360 08:43:35 rpc -- common/autotest_common.sh@10 -- # set +x 00:07:45.360 ************************************ 00:07:45.360 START TEST rpc_trace_cmd_test 00:07:45.360 ************************************ 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1125 -- # rpc_trace_cmd_test 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:07:45.360 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid1653176", 00:07:45.360 "tpoint_group_mask": "0x8", 00:07:45.360 "iscsi_conn": { 00:07:45.360 "mask": "0x2", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "scsi": { 00:07:45.360 "mask": "0x4", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "bdev": { 00:07:45.360 "mask": "0x8", 00:07:45.360 "tpoint_mask": "0xffffffffffffffff" 00:07:45.360 }, 00:07:45.360 "nvmf_rdma": { 00:07:45.360 "mask": "0x10", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "nvmf_tcp": { 00:07:45.360 "mask": "0x20", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "ftl": { 00:07:45.360 "mask": "0x40", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "blobfs": { 00:07:45.360 "mask": "0x80", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "dsa": { 00:07:45.360 "mask": "0x200", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "thread": { 00:07:45.360 "mask": "0x400", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "nvme_pcie": { 00:07:45.360 "mask": "0x800", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "iaa": { 00:07:45.360 "mask": "0x1000", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "nvme_tcp": { 00:07:45.360 "mask": "0x2000", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "bdev_nvme": { 00:07:45.360 "mask": "0x4000", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "sock": { 00:07:45.360 "mask": "0x8000", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "blob": { 00:07:45.360 "mask": "0x10000", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "bdev_raid": { 00:07:45.360 "mask": "0x20000", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 }, 00:07:45.360 "scheduler": { 00:07:45.360 "mask": "0x40000", 00:07:45.360 "tpoint_mask": "0x0" 00:07:45.360 } 00:07:45.360 }' 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 19 -gt 2 ']' 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:07:45.360 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:07:45.621 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:07:45.621 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:07:45.621 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:07:45.621 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:07:45.621 08:43:35 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:07:45.621 00:07:45.621 real 0m0.248s 00:07:45.621 user 0m0.207s 00:07:45.621 sys 0m0.032s 00:07:45.621 08:43:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:45.621 08:43:35 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:07:45.621 ************************************ 00:07:45.621 END TEST rpc_trace_cmd_test 00:07:45.621 ************************************ 00:07:45.621 08:43:35 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:07:45.621 08:43:35 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:07:45.621 08:43:35 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:07:45.621 08:43:35 rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:45.621 08:43:35 rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:45.621 08:43:35 rpc -- common/autotest_common.sh@10 -- # set +x 00:07:45.621 ************************************ 00:07:45.621 START TEST rpc_daemon_integrity 00:07:45.621 ************************************ 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1125 -- # rpc_integrity 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.621 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:07:45.621 { 00:07:45.621 "name": "Malloc2", 00:07:45.621 "aliases": [ 00:07:45.621 "0ec2e5f9-269d-4c4c-a529-c71a11e757ef" 00:07:45.621 ], 00:07:45.621 "product_name": "Malloc disk", 00:07:45.622 "block_size": 512, 00:07:45.622 "num_blocks": 16384, 00:07:45.622 "uuid": "0ec2e5f9-269d-4c4c-a529-c71a11e757ef", 00:07:45.622 "assigned_rate_limits": { 00:07:45.622 "rw_ios_per_sec": 0, 00:07:45.622 "rw_mbytes_per_sec": 0, 00:07:45.622 "r_mbytes_per_sec": 0, 00:07:45.622 "w_mbytes_per_sec": 0 00:07:45.622 }, 00:07:45.622 "claimed": false, 00:07:45.622 "zoned": false, 00:07:45.622 "supported_io_types": { 00:07:45.622 "read": true, 00:07:45.622 "write": true, 00:07:45.622 "unmap": true, 00:07:45.622 "flush": true, 00:07:45.622 "reset": true, 00:07:45.622 "nvme_admin": false, 00:07:45.622 "nvme_io": false, 00:07:45.622 "nvme_io_md": false, 00:07:45.622 "write_zeroes": true, 00:07:45.622 "zcopy": true, 00:07:45.622 "get_zone_info": false, 00:07:45.622 "zone_management": false, 00:07:45.622 "zone_append": false, 00:07:45.622 "compare": false, 00:07:45.622 "compare_and_write": false, 00:07:45.622 "abort": true, 00:07:45.622 "seek_hole": false, 00:07:45.622 "seek_data": false, 00:07:45.622 "copy": true, 00:07:45.622 "nvme_iov_md": false 00:07:45.622 }, 00:07:45.622 "memory_domains": [ 00:07:45.622 { 00:07:45.622 "dma_device_id": "system", 00:07:45.622 "dma_device_type": 1 00:07:45.622 }, 00:07:45.622 { 00:07:45.622 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:45.622 "dma_device_type": 2 00:07:45.622 } 00:07:45.622 ], 00:07:45.622 "driver_specific": {} 00:07:45.622 } 00:07:45.622 ]' 00:07:45.622 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:07:45.926 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:07:45.926 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:07:45.926 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.926 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.926 [2024-11-06 08:43:35.782524] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:07:45.926 [2024-11-06 08:43:35.782556] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:07:45.926 [2024-11-06 08:43:35.782568] vbdev_passthru.c: 681:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0xbda110 00:07:45.927 [2024-11-06 08:43:35.782575] vbdev_passthru.c: 696:vbdev_passthru_register: *NOTICE*: bdev claimed 00:07:45.927 [2024-11-06 08:43:35.783839] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:07:45.927 [2024-11-06 08:43:35.783865] vbdev_passthru.c: 710:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:07:45.927 Passthru0 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:07:45.927 { 00:07:45.927 "name": "Malloc2", 00:07:45.927 "aliases": [ 00:07:45.927 "0ec2e5f9-269d-4c4c-a529-c71a11e757ef" 00:07:45.927 ], 00:07:45.927 "product_name": "Malloc disk", 00:07:45.927 "block_size": 512, 00:07:45.927 "num_blocks": 16384, 00:07:45.927 "uuid": "0ec2e5f9-269d-4c4c-a529-c71a11e757ef", 00:07:45.927 "assigned_rate_limits": { 00:07:45.927 "rw_ios_per_sec": 0, 00:07:45.927 "rw_mbytes_per_sec": 0, 00:07:45.927 "r_mbytes_per_sec": 0, 00:07:45.927 "w_mbytes_per_sec": 0 00:07:45.927 }, 00:07:45.927 "claimed": true, 00:07:45.927 "claim_type": "exclusive_write", 00:07:45.927 "zoned": false, 00:07:45.927 "supported_io_types": { 00:07:45.927 "read": true, 00:07:45.927 "write": true, 00:07:45.927 "unmap": true, 00:07:45.927 "flush": true, 00:07:45.927 "reset": true, 00:07:45.927 "nvme_admin": false, 00:07:45.927 "nvme_io": false, 00:07:45.927 "nvme_io_md": false, 00:07:45.927 "write_zeroes": true, 00:07:45.927 "zcopy": true, 00:07:45.927 "get_zone_info": false, 00:07:45.927 "zone_management": false, 00:07:45.927 "zone_append": false, 00:07:45.927 "compare": false, 00:07:45.927 "compare_and_write": false, 00:07:45.927 "abort": true, 00:07:45.927 "seek_hole": false, 00:07:45.927 "seek_data": false, 00:07:45.927 "copy": true, 00:07:45.927 "nvme_iov_md": false 00:07:45.927 }, 00:07:45.927 "memory_domains": [ 00:07:45.927 { 00:07:45.927 "dma_device_id": "system", 00:07:45.927 "dma_device_type": 1 00:07:45.927 }, 00:07:45.927 { 00:07:45.927 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:45.927 "dma_device_type": 2 00:07:45.927 } 00:07:45.927 ], 00:07:45.927 "driver_specific": {} 00:07:45.927 }, 00:07:45.927 { 00:07:45.927 "name": "Passthru0", 00:07:45.927 "aliases": [ 00:07:45.927 "289d50dd-e90d-5fc5-b5c7-2ce57f604ac5" 00:07:45.927 ], 00:07:45.927 "product_name": "passthru", 00:07:45.927 "block_size": 512, 00:07:45.927 "num_blocks": 16384, 00:07:45.927 "uuid": "289d50dd-e90d-5fc5-b5c7-2ce57f604ac5", 00:07:45.927 "assigned_rate_limits": { 00:07:45.927 "rw_ios_per_sec": 0, 00:07:45.927 "rw_mbytes_per_sec": 0, 00:07:45.927 "r_mbytes_per_sec": 0, 00:07:45.927 "w_mbytes_per_sec": 0 00:07:45.927 }, 00:07:45.927 "claimed": false, 00:07:45.927 "zoned": false, 00:07:45.927 "supported_io_types": { 00:07:45.927 "read": true, 00:07:45.927 "write": true, 00:07:45.927 "unmap": true, 00:07:45.927 "flush": true, 00:07:45.927 "reset": true, 00:07:45.927 "nvme_admin": false, 00:07:45.927 "nvme_io": false, 00:07:45.927 "nvme_io_md": false, 00:07:45.927 "write_zeroes": true, 00:07:45.927 "zcopy": true, 00:07:45.927 "get_zone_info": false, 00:07:45.927 "zone_management": false, 00:07:45.927 "zone_append": false, 00:07:45.927 "compare": false, 00:07:45.927 "compare_and_write": false, 00:07:45.927 "abort": true, 00:07:45.927 "seek_hole": false, 00:07:45.927 "seek_data": false, 00:07:45.927 "copy": true, 00:07:45.927 "nvme_iov_md": false 00:07:45.927 }, 00:07:45.927 "memory_domains": [ 00:07:45.927 { 00:07:45.927 "dma_device_id": "system", 00:07:45.927 "dma_device_type": 1 00:07:45.927 }, 00:07:45.927 { 00:07:45.927 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:07:45.927 "dma_device_type": 2 00:07:45.927 } 00:07:45.927 ], 00:07:45.927 "driver_specific": { 00:07:45.927 "passthru": { 00:07:45.927 "name": "Passthru0", 00:07:45.927 "base_bdev_name": "Malloc2" 00:07:45.927 } 00:07:45.927 } 00:07:45.927 } 00:07:45.927 ]' 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:07:45.927 00:07:45.927 real 0m0.298s 00:07:45.927 user 0m0.193s 00:07:45.927 sys 0m0.040s 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:45.927 08:43:35 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:07:45.927 ************************************ 00:07:45.927 END TEST rpc_daemon_integrity 00:07:45.927 ************************************ 00:07:45.927 08:43:35 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:07:45.927 08:43:35 rpc -- rpc/rpc.sh@84 -- # killprocess 1653176 00:07:45.927 08:43:35 rpc -- common/autotest_common.sh@950 -- # '[' -z 1653176 ']' 00:07:45.927 08:43:35 rpc -- common/autotest_common.sh@954 -- # kill -0 1653176 00:07:45.927 08:43:35 rpc -- common/autotest_common.sh@955 -- # uname 00:07:45.927 08:43:35 rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:45.927 08:43:35 rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1653176 00:07:46.227 08:43:36 rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:46.227 08:43:36 rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:46.227 08:43:36 rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1653176' 00:07:46.227 killing process with pid 1653176 00:07:46.227 08:43:36 rpc -- common/autotest_common.sh@969 -- # kill 1653176 00:07:46.227 08:43:36 rpc -- common/autotest_common.sh@974 -- # wait 1653176 00:07:46.227 00:07:46.227 real 0m2.613s 00:07:46.227 user 0m3.397s 00:07:46.227 sys 0m0.750s 00:07:46.227 08:43:36 rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:46.227 08:43:36 rpc -- common/autotest_common.sh@10 -- # set +x 00:07:46.227 ************************************ 00:07:46.227 END TEST rpc 00:07:46.227 ************************************ 00:07:46.227 08:43:36 -- spdk/autotest.sh@157 -- # run_test skip_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:07:46.227 08:43:36 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:46.227 08:43:36 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:46.227 08:43:36 -- common/autotest_common.sh@10 -- # set +x 00:07:46.496 ************************************ 00:07:46.496 START TEST skip_rpc 00:07:46.496 ************************************ 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:07:46.496 * Looking for test storage... 00:07:46.496 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1689 -- # lcov --version 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@344 -- # case "$op" in 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@345 -- # : 1 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@365 -- # decimal 1 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@353 -- # local d=1 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@355 -- # echo 1 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@366 -- # decimal 2 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@353 -- # local d=2 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@355 -- # echo 2 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:07:46.496 08:43:36 skip_rpc -- scripts/common.sh@368 -- # return 0 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:07:46.496 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:46.496 --rc genhtml_branch_coverage=1 00:07:46.496 --rc genhtml_function_coverage=1 00:07:46.496 --rc genhtml_legend=1 00:07:46.496 --rc geninfo_all_blocks=1 00:07:46.496 --rc geninfo_unexecuted_blocks=1 00:07:46.496 00:07:46.496 ' 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:07:46.496 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:46.496 --rc genhtml_branch_coverage=1 00:07:46.496 --rc genhtml_function_coverage=1 00:07:46.496 --rc genhtml_legend=1 00:07:46.496 --rc geninfo_all_blocks=1 00:07:46.496 --rc geninfo_unexecuted_blocks=1 00:07:46.496 00:07:46.496 ' 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:07:46.496 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:46.496 --rc genhtml_branch_coverage=1 00:07:46.496 --rc genhtml_function_coverage=1 00:07:46.496 --rc genhtml_legend=1 00:07:46.496 --rc geninfo_all_blocks=1 00:07:46.496 --rc geninfo_unexecuted_blocks=1 00:07:46.496 00:07:46.496 ' 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:07:46.496 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:07:46.496 --rc genhtml_branch_coverage=1 00:07:46.496 --rc genhtml_function_coverage=1 00:07:46.496 --rc genhtml_legend=1 00:07:46.496 --rc geninfo_all_blocks=1 00:07:46.496 --rc geninfo_unexecuted_blocks=1 00:07:46.496 00:07:46.496 ' 00:07:46.496 08:43:36 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:07:46.496 08:43:36 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:07:46.496 08:43:36 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:46.496 08:43:36 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:46.496 ************************************ 00:07:46.496 START TEST skip_rpc 00:07:46.496 ************************************ 00:07:46.496 08:43:36 skip_rpc.skip_rpc -- common/autotest_common.sh@1125 -- # test_skip_rpc 00:07:46.496 08:43:36 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=1653803 00:07:46.496 08:43:36 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:07:46.496 08:43:36 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:07:46.496 08:43:36 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:07:46.757 [2024-11-06 08:43:36.613398] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:07:46.757 [2024-11-06 08:43:36.613460] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1653803 ] 00:07:46.757 [2024-11-06 08:43:36.688374] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:46.757 [2024-11-06 08:43:36.730963] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@650 -- # local es=0 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd spdk_get_version 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # rpc_cmd spdk_get_version 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@653 -- # es=1 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 1653803 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@950 -- # '[' -z 1653803 ']' 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # kill -0 1653803 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # uname 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1653803 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1653803' 00:07:52.045 killing process with pid 1653803 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@969 -- # kill 1653803 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@974 -- # wait 1653803 00:07:52.045 00:07:52.045 real 0m5.276s 00:07:52.045 user 0m5.069s 00:07:52.045 sys 0m0.241s 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:52.045 08:43:41 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:52.045 ************************************ 00:07:52.045 END TEST skip_rpc 00:07:52.045 ************************************ 00:07:52.045 08:43:41 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:07:52.045 08:43:41 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:52.045 08:43:41 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:52.045 08:43:41 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:52.045 ************************************ 00:07:52.045 START TEST skip_rpc_with_json 00:07:52.045 ************************************ 00:07:52.045 08:43:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_json 00:07:52.045 08:43:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:07:52.045 08:43:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=1655000 00:07:52.045 08:43:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:07:52.045 08:43:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 1655000 00:07:52.045 08:43:41 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:07:52.045 08:43:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@831 -- # '[' -z 1655000 ']' 00:07:52.045 08:43:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:52.046 08:43:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:52.046 08:43:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:52.046 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:52.046 08:43:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:52.046 08:43:41 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:07:52.046 [2024-11-06 08:43:41.964470] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:07:52.046 [2024-11-06 08:43:41.964524] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1655000 ] 00:07:52.046 [2024-11-06 08:43:42.035765] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:52.046 [2024-11-06 08:43:42.073474] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@864 -- # return 0 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:07:52.989 [2024-11-06 08:43:42.746902] nvmf_rpc.c:2703:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:07:52.989 request: 00:07:52.989 { 00:07:52.989 "trtype": "tcp", 00:07:52.989 "method": "nvmf_get_transports", 00:07:52.989 "req_id": 1 00:07:52.989 } 00:07:52.989 Got JSON-RPC error response 00:07:52.989 response: 00:07:52.989 { 00:07:52.989 "code": -19, 00:07:52.989 "message": "No such device" 00:07:52.989 } 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:07:52.989 [2024-11-06 08:43:42.759029] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:07:52.989 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@561 -- # xtrace_disable 00:07:52.990 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:07:52.990 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:07:52.990 08:43:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:07:52.990 { 00:07:52.990 "subsystems": [ 00:07:52.990 { 00:07:52.990 "subsystem": "fsdev", 00:07:52.990 "config": [ 00:07:52.990 { 00:07:52.990 "method": "fsdev_set_opts", 00:07:52.990 "params": { 00:07:52.990 "fsdev_io_pool_size": 65535, 00:07:52.990 "fsdev_io_cache_size": 256 00:07:52.990 } 00:07:52.990 } 00:07:52.990 ] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "vfio_user_target", 00:07:52.990 "config": null 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "keyring", 00:07:52.990 "config": [] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "iobuf", 00:07:52.990 "config": [ 00:07:52.990 { 00:07:52.990 "method": "iobuf_set_options", 00:07:52.990 "params": { 00:07:52.990 "small_pool_count": 8192, 00:07:52.990 "large_pool_count": 1024, 00:07:52.990 "small_bufsize": 8192, 00:07:52.990 "large_bufsize": 135168, 00:07:52.990 "enable_numa": false 00:07:52.990 } 00:07:52.990 } 00:07:52.990 ] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "sock", 00:07:52.990 "config": [ 00:07:52.990 { 00:07:52.990 "method": "sock_set_default_impl", 00:07:52.990 "params": { 00:07:52.990 "impl_name": "posix" 00:07:52.990 } 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "method": "sock_impl_set_options", 00:07:52.990 "params": { 00:07:52.990 "impl_name": "ssl", 00:07:52.990 "recv_buf_size": 4096, 00:07:52.990 "send_buf_size": 4096, 00:07:52.990 "enable_recv_pipe": true, 00:07:52.990 "enable_quickack": false, 00:07:52.990 "enable_placement_id": 0, 00:07:52.990 "enable_zerocopy_send_server": true, 00:07:52.990 "enable_zerocopy_send_client": false, 00:07:52.990 "zerocopy_threshold": 0, 00:07:52.990 "tls_version": 0, 00:07:52.990 "enable_ktls": false 00:07:52.990 } 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "method": "sock_impl_set_options", 00:07:52.990 "params": { 00:07:52.990 "impl_name": "posix", 00:07:52.990 "recv_buf_size": 2097152, 00:07:52.990 "send_buf_size": 2097152, 00:07:52.990 "enable_recv_pipe": true, 00:07:52.990 "enable_quickack": false, 00:07:52.990 "enable_placement_id": 0, 00:07:52.990 "enable_zerocopy_send_server": true, 00:07:52.990 "enable_zerocopy_send_client": false, 00:07:52.990 "zerocopy_threshold": 0, 00:07:52.990 "tls_version": 0, 00:07:52.990 "enable_ktls": false 00:07:52.990 } 00:07:52.990 } 00:07:52.990 ] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "vmd", 00:07:52.990 "config": [] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "accel", 00:07:52.990 "config": [ 00:07:52.990 { 00:07:52.990 "method": "accel_set_options", 00:07:52.990 "params": { 00:07:52.990 "small_cache_size": 128, 00:07:52.990 "large_cache_size": 16, 00:07:52.990 "task_count": 2048, 00:07:52.990 "sequence_count": 2048, 00:07:52.990 "buf_count": 2048 00:07:52.990 } 00:07:52.990 } 00:07:52.990 ] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "bdev", 00:07:52.990 "config": [ 00:07:52.990 { 00:07:52.990 "method": "bdev_set_options", 00:07:52.990 "params": { 00:07:52.990 "bdev_io_pool_size": 65535, 00:07:52.990 "bdev_io_cache_size": 256, 00:07:52.990 "bdev_auto_examine": true, 00:07:52.990 "iobuf_small_cache_size": 128, 00:07:52.990 "iobuf_large_cache_size": 16 00:07:52.990 } 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "method": "bdev_raid_set_options", 00:07:52.990 "params": { 00:07:52.990 "process_window_size_kb": 1024, 00:07:52.990 "process_max_bandwidth_mb_sec": 0 00:07:52.990 } 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "method": "bdev_iscsi_set_options", 00:07:52.990 "params": { 00:07:52.990 "timeout_sec": 30 00:07:52.990 } 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "method": "bdev_nvme_set_options", 00:07:52.990 "params": { 00:07:52.990 "action_on_timeout": "none", 00:07:52.990 "timeout_us": 0, 00:07:52.990 "timeout_admin_us": 0, 00:07:52.990 "keep_alive_timeout_ms": 10000, 00:07:52.990 "arbitration_burst": 0, 00:07:52.990 "low_priority_weight": 0, 00:07:52.990 "medium_priority_weight": 0, 00:07:52.990 "high_priority_weight": 0, 00:07:52.990 "nvme_adminq_poll_period_us": 10000, 00:07:52.990 "nvme_ioq_poll_period_us": 0, 00:07:52.990 "io_queue_requests": 0, 00:07:52.990 "delay_cmd_submit": true, 00:07:52.990 "transport_retry_count": 4, 00:07:52.990 "bdev_retry_count": 3, 00:07:52.990 "transport_ack_timeout": 0, 00:07:52.990 "ctrlr_loss_timeout_sec": 0, 00:07:52.990 "reconnect_delay_sec": 0, 00:07:52.990 "fast_io_fail_timeout_sec": 0, 00:07:52.990 "disable_auto_failback": false, 00:07:52.990 "generate_uuids": false, 00:07:52.990 "transport_tos": 0, 00:07:52.990 "nvme_error_stat": false, 00:07:52.990 "rdma_srq_size": 0, 00:07:52.990 "io_path_stat": false, 00:07:52.990 "allow_accel_sequence": false, 00:07:52.990 "rdma_max_cq_size": 0, 00:07:52.990 "rdma_cm_event_timeout_ms": 0, 00:07:52.990 "dhchap_digests": [ 00:07:52.990 "sha256", 00:07:52.990 "sha384", 00:07:52.990 "sha512" 00:07:52.990 ], 00:07:52.990 "dhchap_dhgroups": [ 00:07:52.990 "null", 00:07:52.990 "ffdhe2048", 00:07:52.990 "ffdhe3072", 00:07:52.990 "ffdhe4096", 00:07:52.990 "ffdhe6144", 00:07:52.990 "ffdhe8192" 00:07:52.990 ] 00:07:52.990 } 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "method": "bdev_nvme_set_hotplug", 00:07:52.990 "params": { 00:07:52.990 "period_us": 100000, 00:07:52.990 "enable": false 00:07:52.990 } 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "method": "bdev_wait_for_examine" 00:07:52.990 } 00:07:52.990 ] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "scsi", 00:07:52.990 "config": null 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "scheduler", 00:07:52.990 "config": [ 00:07:52.990 { 00:07:52.990 "method": "framework_set_scheduler", 00:07:52.990 "params": { 00:07:52.990 "name": "static" 00:07:52.990 } 00:07:52.990 } 00:07:52.990 ] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "vhost_scsi", 00:07:52.990 "config": [] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "vhost_blk", 00:07:52.990 "config": [] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "ublk", 00:07:52.990 "config": [] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "nbd", 00:07:52.990 "config": [] 00:07:52.990 }, 00:07:52.990 { 00:07:52.990 "subsystem": "nvmf", 00:07:52.990 "config": [ 00:07:52.990 { 00:07:52.990 "method": "nvmf_set_config", 00:07:52.990 "params": { 00:07:52.990 "discovery_filter": "match_any", 00:07:52.990 "admin_cmd_passthru": { 00:07:52.990 "identify_ctrlr": false 00:07:52.990 }, 00:07:52.990 "dhchap_digests": [ 00:07:52.991 "sha256", 00:07:52.991 "sha384", 00:07:52.991 "sha512" 00:07:52.991 ], 00:07:52.991 "dhchap_dhgroups": [ 00:07:52.991 "null", 00:07:52.991 "ffdhe2048", 00:07:52.991 "ffdhe3072", 00:07:52.991 "ffdhe4096", 00:07:52.991 "ffdhe6144", 00:07:52.991 "ffdhe8192" 00:07:52.991 ] 00:07:52.991 } 00:07:52.991 }, 00:07:52.991 { 00:07:52.991 "method": "nvmf_set_max_subsystems", 00:07:52.991 "params": { 00:07:52.991 "max_subsystems": 1024 00:07:52.991 } 00:07:52.991 }, 00:07:52.991 { 00:07:52.991 "method": "nvmf_set_crdt", 00:07:52.991 "params": { 00:07:52.991 "crdt1": 0, 00:07:52.991 "crdt2": 0, 00:07:52.991 "crdt3": 0 00:07:52.991 } 00:07:52.991 }, 00:07:52.991 { 00:07:52.991 "method": "nvmf_create_transport", 00:07:52.991 "params": { 00:07:52.991 "trtype": "TCP", 00:07:52.991 "max_queue_depth": 128, 00:07:52.991 "max_io_qpairs_per_ctrlr": 127, 00:07:52.991 "in_capsule_data_size": 4096, 00:07:52.991 "max_io_size": 131072, 00:07:52.991 "io_unit_size": 131072, 00:07:52.991 "max_aq_depth": 128, 00:07:52.991 "num_shared_buffers": 511, 00:07:52.991 "buf_cache_size": 4294967295, 00:07:52.991 "dif_insert_or_strip": false, 00:07:52.991 "zcopy": false, 00:07:52.991 "c2h_success": true, 00:07:52.991 "sock_priority": 0, 00:07:52.991 "abort_timeout_sec": 1, 00:07:52.991 "ack_timeout": 0, 00:07:52.991 "data_wr_pool_size": 0 00:07:52.991 } 00:07:52.991 } 00:07:52.991 ] 00:07:52.991 }, 00:07:52.991 { 00:07:52.991 "subsystem": "iscsi", 00:07:52.991 "config": [ 00:07:52.991 { 00:07:52.991 "method": "iscsi_set_options", 00:07:52.991 "params": { 00:07:52.991 "node_base": "iqn.2016-06.io.spdk", 00:07:52.991 "max_sessions": 128, 00:07:52.991 "max_connections_per_session": 2, 00:07:52.991 "max_queue_depth": 64, 00:07:52.991 "default_time2wait": 2, 00:07:52.991 "default_time2retain": 20, 00:07:52.991 "first_burst_length": 8192, 00:07:52.991 "immediate_data": true, 00:07:52.991 "allow_duplicated_isid": false, 00:07:52.991 "error_recovery_level": 0, 00:07:52.991 "nop_timeout": 60, 00:07:52.991 "nop_in_interval": 30, 00:07:52.991 "disable_chap": false, 00:07:52.991 "require_chap": false, 00:07:52.991 "mutual_chap": false, 00:07:52.991 "chap_group": 0, 00:07:52.991 "max_large_datain_per_connection": 64, 00:07:52.991 "max_r2t_per_connection": 4, 00:07:52.991 "pdu_pool_size": 36864, 00:07:52.991 "immediate_data_pool_size": 16384, 00:07:52.991 "data_out_pool_size": 2048 00:07:52.991 } 00:07:52.991 } 00:07:52.991 ] 00:07:52.991 } 00:07:52.991 ] 00:07:52.991 } 00:07:52.991 08:43:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:07:52.991 08:43:42 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 1655000 00:07:52.991 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 1655000 ']' 00:07:52.991 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 1655000 00:07:52.991 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:07:52.991 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:52.991 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1655000 00:07:52.991 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:52.991 08:43:42 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:52.991 08:43:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1655000' 00:07:52.991 killing process with pid 1655000 00:07:52.991 08:43:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 1655000 00:07:52.991 08:43:43 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 1655000 00:07:53.252 08:43:43 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=1655148 00:07:53.252 08:43:43 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:07:53.252 08:43:43 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 1655148 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@950 -- # '[' -z 1655148 ']' 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # kill -0 1655148 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # uname 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1655148 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1655148' 00:07:58.544 killing process with pid 1655148 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@969 -- # kill 1655148 00:07:58.544 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@974 -- # wait 1655148 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:07:58.545 00:07:58.545 real 0m6.588s 00:07:58.545 user 0m6.464s 00:07:58.545 sys 0m0.579s 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:07:58.545 ************************************ 00:07:58.545 END TEST skip_rpc_with_json 00:07:58.545 ************************************ 00:07:58.545 08:43:48 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:07:58.545 08:43:48 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:58.545 08:43:48 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:58.545 08:43:48 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:58.545 ************************************ 00:07:58.545 START TEST skip_rpc_with_delay 00:07:58.545 ************************************ 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1125 -- # test_skip_rpc_with_delay 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@650 -- # local es=0 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:07:58.545 [2024-11-06 08:43:48.638000] app.c: 842:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@653 -- # es=1 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:58.545 00:07:58.545 real 0m0.079s 00:07:58.545 user 0m0.048s 00:07:58.545 sys 0m0.031s 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1126 -- # xtrace_disable 00:07:58.545 08:43:48 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:07:58.545 ************************************ 00:07:58.545 END TEST skip_rpc_with_delay 00:07:58.545 ************************************ 00:07:58.805 08:43:48 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:07:58.805 08:43:48 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:07:58.805 08:43:48 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:07:58.805 08:43:48 skip_rpc -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:07:58.805 08:43:48 skip_rpc -- common/autotest_common.sh@1107 -- # xtrace_disable 00:07:58.805 08:43:48 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:07:58.805 ************************************ 00:07:58.805 START TEST exit_on_failed_rpc_init 00:07:58.805 ************************************ 00:07:58.805 08:43:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1125 -- # test_exit_on_failed_rpc_init 00:07:58.805 08:43:48 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=1656464 00:07:58.805 08:43:48 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 1656464 00:07:58.805 08:43:48 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:07:58.805 08:43:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@831 -- # '[' -z 1656464 ']' 00:07:58.805 08:43:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:58.805 08:43:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@836 -- # local max_retries=100 00:07:58.805 08:43:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:58.805 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:58.805 08:43:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@840 -- # xtrace_disable 00:07:58.805 08:43:48 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:07:58.805 [2024-11-06 08:43:48.797725] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:07:58.805 [2024-11-06 08:43:48.797799] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1656464 ] 00:07:58.805 [2024-11-06 08:43:48.871874] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:58.805 [2024-11-06 08:43:48.913807] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@864 -- # return 0 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@650 -- # local es=0 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:07:59.744 [2024-11-06 08:43:49.648095] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:07:59.744 [2024-11-06 08:43:49.648149] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1656498 ] 00:07:59.744 [2024-11-06 08:43:49.733328] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:07:59.744 [2024-11-06 08:43:49.769050] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:07:59.744 [2024-11-06 08:43:49.769103] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:07:59.744 [2024-11-06 08:43:49.769112] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:07:59.744 [2024-11-06 08:43:49.769119] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@653 -- # es=234 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@662 -- # es=106 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@663 -- # case "$es" in 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@670 -- # es=1 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 1656464 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@950 -- # '[' -z 1656464 ']' 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # kill -0 1656464 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # uname 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:07:59.744 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1656464 00:08:00.004 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:00.004 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:00.004 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1656464' 00:08:00.004 killing process with pid 1656464 00:08:00.004 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@969 -- # kill 1656464 00:08:00.004 08:43:49 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@974 -- # wait 1656464 00:08:00.004 00:08:00.004 real 0m1.346s 00:08:00.004 user 0m1.574s 00:08:00.004 sys 0m0.382s 00:08:00.004 08:43:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:00.004 08:43:50 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:08:00.004 ************************************ 00:08:00.004 END TEST exit_on_failed_rpc_init 00:08:00.004 ************************************ 00:08:00.269 08:43:50 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:08:00.270 00:08:00.270 real 0m13.799s 00:08:00.270 user 0m13.364s 00:08:00.270 sys 0m1.559s 00:08:00.270 08:43:50 skip_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:00.270 08:43:50 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:00.270 ************************************ 00:08:00.270 END TEST skip_rpc 00:08:00.270 ************************************ 00:08:00.270 08:43:50 -- spdk/autotest.sh@158 -- # run_test rpc_client /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:08:00.270 08:43:50 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:00.270 08:43:50 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:00.270 08:43:50 -- common/autotest_common.sh@10 -- # set +x 00:08:00.270 ************************************ 00:08:00.270 START TEST rpc_client 00:08:00.270 ************************************ 00:08:00.270 08:43:50 rpc_client -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:08:00.270 * Looking for test storage... 00:08:00.270 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client 00:08:00.270 08:43:50 rpc_client -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:00.270 08:43:50 rpc_client -- common/autotest_common.sh@1689 -- # lcov --version 00:08:00.270 08:43:50 rpc_client -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:00.530 08:43:50 rpc_client -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@336 -- # IFS=.-: 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@336 -- # read -ra ver1 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@337 -- # IFS=.-: 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@337 -- # read -ra ver2 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@338 -- # local 'op=<' 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@340 -- # ver1_l=2 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@341 -- # ver2_l=1 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@344 -- # case "$op" in 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@345 -- # : 1 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@365 -- # decimal 1 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@353 -- # local d=1 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@355 -- # echo 1 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@365 -- # ver1[v]=1 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@366 -- # decimal 2 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@353 -- # local d=2 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@355 -- # echo 2 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@366 -- # ver2[v]=2 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:00.530 08:43:50 rpc_client -- scripts/common.sh@368 -- # return 0 00:08:00.530 08:43:50 rpc_client -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:00.530 08:43:50 rpc_client -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:00.530 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:00.530 --rc genhtml_branch_coverage=1 00:08:00.530 --rc genhtml_function_coverage=1 00:08:00.530 --rc genhtml_legend=1 00:08:00.530 --rc geninfo_all_blocks=1 00:08:00.530 --rc geninfo_unexecuted_blocks=1 00:08:00.530 00:08:00.530 ' 00:08:00.530 08:43:50 rpc_client -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:00.530 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:00.530 --rc genhtml_branch_coverage=1 00:08:00.530 --rc genhtml_function_coverage=1 00:08:00.530 --rc genhtml_legend=1 00:08:00.530 --rc geninfo_all_blocks=1 00:08:00.530 --rc geninfo_unexecuted_blocks=1 00:08:00.530 00:08:00.530 ' 00:08:00.530 08:43:50 rpc_client -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:00.530 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:00.530 --rc genhtml_branch_coverage=1 00:08:00.530 --rc genhtml_function_coverage=1 00:08:00.530 --rc genhtml_legend=1 00:08:00.530 --rc geninfo_all_blocks=1 00:08:00.530 --rc geninfo_unexecuted_blocks=1 00:08:00.530 00:08:00.530 ' 00:08:00.530 08:43:50 rpc_client -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:00.530 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:00.530 --rc genhtml_branch_coverage=1 00:08:00.530 --rc genhtml_function_coverage=1 00:08:00.530 --rc genhtml_legend=1 00:08:00.530 --rc geninfo_all_blocks=1 00:08:00.530 --rc geninfo_unexecuted_blocks=1 00:08:00.530 00:08:00.530 ' 00:08:00.531 08:43:50 rpc_client -- rpc_client/rpc_client.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client_test 00:08:00.531 OK 00:08:00.531 08:43:50 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:08:00.531 00:08:00.531 real 0m0.230s 00:08:00.531 user 0m0.138s 00:08:00.531 sys 0m0.104s 00:08:00.531 08:43:50 rpc_client -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:00.531 08:43:50 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:08:00.531 ************************************ 00:08:00.531 END TEST rpc_client 00:08:00.531 ************************************ 00:08:00.531 08:43:50 -- spdk/autotest.sh@159 -- # run_test json_config /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:08:00.531 08:43:50 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:00.531 08:43:50 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:00.531 08:43:50 -- common/autotest_common.sh@10 -- # set +x 00:08:00.531 ************************************ 00:08:00.531 START TEST json_config 00:08:00.531 ************************************ 00:08:00.531 08:43:50 json_config -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:08:00.531 08:43:50 json_config -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:00.531 08:43:50 json_config -- common/autotest_common.sh@1689 -- # lcov --version 00:08:00.531 08:43:50 json_config -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:00.793 08:43:50 json_config -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:00.793 08:43:50 json_config -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:00.793 08:43:50 json_config -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:00.793 08:43:50 json_config -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:00.793 08:43:50 json_config -- scripts/common.sh@336 -- # IFS=.-: 00:08:00.793 08:43:50 json_config -- scripts/common.sh@336 -- # read -ra ver1 00:08:00.793 08:43:50 json_config -- scripts/common.sh@337 -- # IFS=.-: 00:08:00.793 08:43:50 json_config -- scripts/common.sh@337 -- # read -ra ver2 00:08:00.793 08:43:50 json_config -- scripts/common.sh@338 -- # local 'op=<' 00:08:00.793 08:43:50 json_config -- scripts/common.sh@340 -- # ver1_l=2 00:08:00.793 08:43:50 json_config -- scripts/common.sh@341 -- # ver2_l=1 00:08:00.793 08:43:50 json_config -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:00.793 08:43:50 json_config -- scripts/common.sh@344 -- # case "$op" in 00:08:00.793 08:43:50 json_config -- scripts/common.sh@345 -- # : 1 00:08:00.793 08:43:50 json_config -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:00.793 08:43:50 json_config -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:00.793 08:43:50 json_config -- scripts/common.sh@365 -- # decimal 1 00:08:00.793 08:43:50 json_config -- scripts/common.sh@353 -- # local d=1 00:08:00.793 08:43:50 json_config -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:00.793 08:43:50 json_config -- scripts/common.sh@355 -- # echo 1 00:08:00.793 08:43:50 json_config -- scripts/common.sh@365 -- # ver1[v]=1 00:08:00.793 08:43:50 json_config -- scripts/common.sh@366 -- # decimal 2 00:08:00.793 08:43:50 json_config -- scripts/common.sh@353 -- # local d=2 00:08:00.793 08:43:50 json_config -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:00.793 08:43:50 json_config -- scripts/common.sh@355 -- # echo 2 00:08:00.793 08:43:50 json_config -- scripts/common.sh@366 -- # ver2[v]=2 00:08:00.793 08:43:50 json_config -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:00.793 08:43:50 json_config -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:00.793 08:43:50 json_config -- scripts/common.sh@368 -- # return 0 00:08:00.793 08:43:50 json_config -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:00.793 08:43:50 json_config -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:00.793 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:00.793 --rc genhtml_branch_coverage=1 00:08:00.793 --rc genhtml_function_coverage=1 00:08:00.793 --rc genhtml_legend=1 00:08:00.793 --rc geninfo_all_blocks=1 00:08:00.793 --rc geninfo_unexecuted_blocks=1 00:08:00.793 00:08:00.793 ' 00:08:00.793 08:43:50 json_config -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:00.793 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:00.793 --rc genhtml_branch_coverage=1 00:08:00.793 --rc genhtml_function_coverage=1 00:08:00.793 --rc genhtml_legend=1 00:08:00.793 --rc geninfo_all_blocks=1 00:08:00.793 --rc geninfo_unexecuted_blocks=1 00:08:00.793 00:08:00.793 ' 00:08:00.793 08:43:50 json_config -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:00.793 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:00.793 --rc genhtml_branch_coverage=1 00:08:00.793 --rc genhtml_function_coverage=1 00:08:00.793 --rc genhtml_legend=1 00:08:00.793 --rc geninfo_all_blocks=1 00:08:00.793 --rc geninfo_unexecuted_blocks=1 00:08:00.793 00:08:00.793 ' 00:08:00.793 08:43:50 json_config -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:00.793 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:00.793 --rc genhtml_branch_coverage=1 00:08:00.793 --rc genhtml_function_coverage=1 00:08:00.793 --rc genhtml_legend=1 00:08:00.793 --rc geninfo_all_blocks=1 00:08:00.793 --rc geninfo_unexecuted_blocks=1 00:08:00.793 00:08:00.793 ' 00:08:00.793 08:43:50 json_config -- json_config/json_config.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@7 -- # uname -s 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:00.793 08:43:50 json_config -- scripts/common.sh@15 -- # shopt -s extglob 00:08:00.793 08:43:50 json_config -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:00.793 08:43:50 json_config -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:00.793 08:43:50 json_config -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:00.793 08:43:50 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:00.793 08:43:50 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:00.793 08:43:50 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:00.793 08:43:50 json_config -- paths/export.sh@5 -- # export PATH 00:08:00.793 08:43:50 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@51 -- # : 0 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:08:00.793 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:08:00.793 08:43:50 json_config -- nvmf/common.sh@55 -- # have_pci_nics=0 00:08:00.793 08:43:50 json_config -- json_config/json_config.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:08:00.793 08:43:50 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:08:00.793 08:43:50 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:08:00.793 08:43:50 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:08:00.793 08:43:50 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:08:00.793 08:43:50 json_config -- json_config/json_config.sh@31 -- # app_pid=(['target']='' ['initiator']='') 00:08:00.793 08:43:50 json_config -- json_config/json_config.sh@31 -- # declare -A app_pid 00:08:00.793 08:43:50 json_config -- json_config/json_config.sh@32 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock' ['initiator']='/var/tmp/spdk_initiator.sock') 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@32 -- # declare -A app_socket 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@33 -- # app_params=(['target']='-m 0x1 -s 1024' ['initiator']='-m 0x2 -g -u -s 1024') 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@33 -- # declare -A app_params 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@34 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json' ['initiator']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json') 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@34 -- # declare -A configs_path 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@40 -- # last_event_id=0 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@362 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@363 -- # echo 'INFO: JSON configuration test init' 00:08:00.794 INFO: JSON configuration test init 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@364 -- # json_config_test_init 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@269 -- # timing_enter json_config_test_init 00:08:00.794 08:43:50 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:08:00.794 08:43:50 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@270 -- # timing_enter json_config_setup_target 00:08:00.794 08:43:50 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:08:00.794 08:43:50 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:00.794 08:43:50 json_config -- json_config/json_config.sh@272 -- # json_config_test_start_app target --wait-for-rpc 00:08:00.794 08:43:50 json_config -- json_config/common.sh@9 -- # local app=target 00:08:00.794 08:43:50 json_config -- json_config/common.sh@10 -- # shift 00:08:00.794 08:43:50 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:08:00.794 08:43:50 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:08:00.794 08:43:50 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:08:00.794 08:43:50 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:08:00.794 08:43:50 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:08:00.794 08:43:50 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=1656952 00:08:00.794 08:43:50 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:08:00.794 Waiting for target to run... 00:08:00.794 08:43:50 json_config -- json_config/common.sh@25 -- # waitforlisten 1656952 /var/tmp/spdk_tgt.sock 00:08:00.794 08:43:50 json_config -- common/autotest_common.sh@831 -- # '[' -z 1656952 ']' 00:08:00.794 08:43:50 json_config -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:08:00.794 08:43:50 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --wait-for-rpc 00:08:00.794 08:43:50 json_config -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:00.794 08:43:50 json_config -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:08:00.794 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:08:00.794 08:43:50 json_config -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:00.794 08:43:50 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:00.794 [2024-11-06 08:43:50.781092] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:00.794 [2024-11-06 08:43:50.781166] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1656952 ] 00:08:01.054 [2024-11-06 08:43:51.113224] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:01.054 [2024-11-06 08:43:51.142837] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:01.626 08:43:51 json_config -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:01.626 08:43:51 json_config -- common/autotest_common.sh@864 -- # return 0 00:08:01.626 08:43:51 json_config -- json_config/common.sh@26 -- # echo '' 00:08:01.626 00:08:01.626 08:43:51 json_config -- json_config/json_config.sh@276 -- # create_accel_config 00:08:01.626 08:43:51 json_config -- json_config/json_config.sh@100 -- # timing_enter create_accel_config 00:08:01.626 08:43:51 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:08:01.626 08:43:51 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:01.626 08:43:51 json_config -- json_config/json_config.sh@102 -- # [[ 0 -eq 1 ]] 00:08:01.626 08:43:51 json_config -- json_config/json_config.sh@108 -- # timing_exit create_accel_config 00:08:01.626 08:43:51 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:01.626 08:43:51 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:01.626 08:43:51 json_config -- json_config/json_config.sh@280 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh --json-with-subsystems 00:08:01.626 08:43:51 json_config -- json_config/json_config.sh@281 -- # tgt_rpc load_config 00:08:01.626 08:43:51 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock load_config 00:08:02.197 08:43:52 json_config -- json_config/json_config.sh@283 -- # tgt_check_notification_types 00:08:02.197 08:43:52 json_config -- json_config/json_config.sh@43 -- # timing_enter tgt_check_notification_types 00:08:02.197 08:43:52 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:08:02.197 08:43:52 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:02.197 08:43:52 json_config -- json_config/json_config.sh@45 -- # local ret=0 00:08:02.197 08:43:52 json_config -- json_config/json_config.sh@46 -- # enabled_types=('bdev_register' 'bdev_unregister') 00:08:02.197 08:43:52 json_config -- json_config/json_config.sh@46 -- # local enabled_types 00:08:02.197 08:43:52 json_config -- json_config/json_config.sh@47 -- # [[ y == y ]] 00:08:02.197 08:43:52 json_config -- json_config/json_config.sh@48 -- # enabled_types+=("fsdev_register" "fsdev_unregister") 00:08:02.197 08:43:52 json_config -- json_config/json_config.sh@51 -- # tgt_rpc notify_get_types 00:08:02.197 08:43:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock notify_get_types 00:08:02.197 08:43:52 json_config -- json_config/json_config.sh@51 -- # jq -r '.[]' 00:08:02.456 08:43:52 json_config -- json_config/json_config.sh@51 -- # get_types=('fsdev_register' 'fsdev_unregister' 'bdev_register' 'bdev_unregister') 00:08:02.456 08:43:52 json_config -- json_config/json_config.sh@51 -- # local get_types 00:08:02.456 08:43:52 json_config -- json_config/json_config.sh@53 -- # local type_diff 00:08:02.456 08:43:52 json_config -- json_config/json_config.sh@54 -- # echo bdev_register bdev_unregister fsdev_register fsdev_unregister fsdev_register fsdev_unregister bdev_register bdev_unregister 00:08:02.456 08:43:52 json_config -- json_config/json_config.sh@54 -- # tr ' ' '\n' 00:08:02.456 08:43:52 json_config -- json_config/json_config.sh@54 -- # sort 00:08:02.456 08:43:52 json_config -- json_config/json_config.sh@54 -- # uniq -u 00:08:02.456 08:43:52 json_config -- json_config/json_config.sh@54 -- # type_diff= 00:08:02.456 08:43:52 json_config -- json_config/json_config.sh@56 -- # [[ -n '' ]] 00:08:02.456 08:43:52 json_config -- json_config/json_config.sh@61 -- # timing_exit tgt_check_notification_types 00:08:02.456 08:43:52 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:02.457 08:43:52 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@62 -- # return 0 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@285 -- # [[ 0 -eq 1 ]] 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@289 -- # [[ 0 -eq 1 ]] 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@293 -- # [[ 0 -eq 1 ]] 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@297 -- # [[ 1 -eq 1 ]] 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@298 -- # create_nvmf_subsystem_config 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@237 -- # timing_enter create_nvmf_subsystem_config 00:08:02.457 08:43:52 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:08:02.457 08:43:52 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@239 -- # NVMF_FIRST_TARGET_IP=127.0.0.1 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@240 -- # [[ tcp == \r\d\m\a ]] 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@244 -- # [[ -z 127.0.0.1 ]] 00:08:02.457 08:43:52 json_config -- json_config/json_config.sh@249 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocForNvmf0 00:08:02.457 08:43:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocForNvmf0 00:08:02.718 MallocForNvmf0 00:08:02.718 08:43:52 json_config -- json_config/json_config.sh@250 -- # tgt_rpc bdev_malloc_create 4 1024 --name MallocForNvmf1 00:08:02.718 08:43:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 4 1024 --name MallocForNvmf1 00:08:02.718 MallocForNvmf1 00:08:02.718 08:43:52 json_config -- json_config/json_config.sh@252 -- # tgt_rpc nvmf_create_transport -t tcp -u 8192 -c 0 00:08:02.718 08:43:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_transport -t tcp -u 8192 -c 0 00:08:02.979 [2024-11-06 08:43:52.963547] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:08:02.979 08:43:52 json_config -- json_config/json_config.sh@253 -- # tgt_rpc nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:08:02.979 08:43:52 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:08:03.239 08:43:53 json_config -- json_config/json_config.sh@254 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:08:03.239 08:43:53 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:08:03.239 08:43:53 json_config -- json_config/json_config.sh@255 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:08:03.239 08:43:53 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:08:03.500 08:43:53 json_config -- json_config/json_config.sh@256 -- # tgt_rpc nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:08:03.500 08:43:53 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:08:03.761 [2024-11-06 08:43:53.677812] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:08:03.761 08:43:53 json_config -- json_config/json_config.sh@258 -- # timing_exit create_nvmf_subsystem_config 00:08:03.761 08:43:53 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:03.761 08:43:53 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:03.761 08:43:53 json_config -- json_config/json_config.sh@300 -- # timing_exit json_config_setup_target 00:08:03.761 08:43:53 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:03.761 08:43:53 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:03.761 08:43:53 json_config -- json_config/json_config.sh@302 -- # [[ 0 -eq 1 ]] 00:08:03.761 08:43:53 json_config -- json_config/json_config.sh@307 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:08:03.761 08:43:53 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:08:04.021 MallocBdevForConfigChangeCheck 00:08:04.021 08:43:53 json_config -- json_config/json_config.sh@309 -- # timing_exit json_config_test_init 00:08:04.021 08:43:53 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:04.021 08:43:53 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:04.021 08:43:54 json_config -- json_config/json_config.sh@366 -- # tgt_rpc save_config 00:08:04.021 08:43:54 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:08:04.282 08:43:54 json_config -- json_config/json_config.sh@368 -- # echo 'INFO: shutting down applications...' 00:08:04.282 INFO: shutting down applications... 00:08:04.282 08:43:54 json_config -- json_config/json_config.sh@369 -- # [[ 0 -eq 1 ]] 00:08:04.282 08:43:54 json_config -- json_config/json_config.sh@375 -- # json_config_clear target 00:08:04.282 08:43:54 json_config -- json_config/json_config.sh@339 -- # [[ -n 22 ]] 00:08:04.282 08:43:54 json_config -- json_config/json_config.sh@340 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py -s /var/tmp/spdk_tgt.sock clear_config 00:08:04.853 Calling clear_iscsi_subsystem 00:08:04.853 Calling clear_nvmf_subsystem 00:08:04.853 Calling clear_nbd_subsystem 00:08:04.853 Calling clear_ublk_subsystem 00:08:04.853 Calling clear_vhost_blk_subsystem 00:08:04.853 Calling clear_vhost_scsi_subsystem 00:08:04.853 Calling clear_bdev_subsystem 00:08:04.853 08:43:54 json_config -- json_config/json_config.sh@344 -- # local config_filter=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py 00:08:04.853 08:43:54 json_config -- json_config/json_config.sh@350 -- # count=100 00:08:04.853 08:43:54 json_config -- json_config/json_config.sh@351 -- # '[' 100 -gt 0 ']' 00:08:04.853 08:43:54 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:08:04.853 08:43:54 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method delete_global_parameters 00:08:04.853 08:43:54 json_config -- json_config/json_config.sh@352 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method check_empty 00:08:05.113 08:43:55 json_config -- json_config/json_config.sh@352 -- # break 00:08:05.113 08:43:55 json_config -- json_config/json_config.sh@357 -- # '[' 100 -eq 0 ']' 00:08:05.113 08:43:55 json_config -- json_config/json_config.sh@376 -- # json_config_test_shutdown_app target 00:08:05.113 08:43:55 json_config -- json_config/common.sh@31 -- # local app=target 00:08:05.113 08:43:55 json_config -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:08:05.113 08:43:55 json_config -- json_config/common.sh@35 -- # [[ -n 1656952 ]] 00:08:05.113 08:43:55 json_config -- json_config/common.sh@38 -- # kill -SIGINT 1656952 00:08:05.113 08:43:55 json_config -- json_config/common.sh@40 -- # (( i = 0 )) 00:08:05.113 08:43:55 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:08:05.113 08:43:55 json_config -- json_config/common.sh@41 -- # kill -0 1656952 00:08:05.113 08:43:55 json_config -- json_config/common.sh@45 -- # sleep 0.5 00:08:05.684 08:43:55 json_config -- json_config/common.sh@40 -- # (( i++ )) 00:08:05.684 08:43:55 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:08:05.684 08:43:55 json_config -- json_config/common.sh@41 -- # kill -0 1656952 00:08:05.684 08:43:55 json_config -- json_config/common.sh@42 -- # app_pid["$app"]= 00:08:05.684 08:43:55 json_config -- json_config/common.sh@43 -- # break 00:08:05.684 08:43:55 json_config -- json_config/common.sh@48 -- # [[ -n '' ]] 00:08:05.684 08:43:55 json_config -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:08:05.684 SPDK target shutdown done 00:08:05.684 08:43:55 json_config -- json_config/json_config.sh@378 -- # echo 'INFO: relaunching applications...' 00:08:05.684 INFO: relaunching applications... 00:08:05.684 08:43:55 json_config -- json_config/json_config.sh@379 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:05.684 08:43:55 json_config -- json_config/common.sh@9 -- # local app=target 00:08:05.684 08:43:55 json_config -- json_config/common.sh@10 -- # shift 00:08:05.684 08:43:55 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:08:05.684 08:43:55 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:08:05.684 08:43:55 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:08:05.684 08:43:55 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:08:05.684 08:43:55 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:08:05.684 08:43:55 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=1658091 00:08:05.684 08:43:55 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:08:05.684 Waiting for target to run... 00:08:05.684 08:43:55 json_config -- json_config/common.sh@25 -- # waitforlisten 1658091 /var/tmp/spdk_tgt.sock 00:08:05.684 08:43:55 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:05.684 08:43:55 json_config -- common/autotest_common.sh@831 -- # '[' -z 1658091 ']' 00:08:05.684 08:43:55 json_config -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:08:05.684 08:43:55 json_config -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:05.684 08:43:55 json_config -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:08:05.684 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:08:05.684 08:43:55 json_config -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:05.685 08:43:55 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:05.685 [2024-11-06 08:43:55.645742] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:05.685 [2024-11-06 08:43:55.645819] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1658091 ] 00:08:05.945 [2024-11-06 08:43:55.929814] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:05.945 [2024-11-06 08:43:55.959242] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:06.515 [2024-11-06 08:43:56.477202] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:08:06.515 [2024-11-06 08:43:56.509582] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:08:06.515 08:43:56 json_config -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:06.515 08:43:56 json_config -- common/autotest_common.sh@864 -- # return 0 00:08:06.515 08:43:56 json_config -- json_config/common.sh@26 -- # echo '' 00:08:06.515 00:08:06.515 08:43:56 json_config -- json_config/json_config.sh@380 -- # [[ 0 -eq 1 ]] 00:08:06.515 08:43:56 json_config -- json_config/json_config.sh@384 -- # echo 'INFO: Checking if target configuration is the same...' 00:08:06.515 INFO: Checking if target configuration is the same... 00:08:06.515 08:43:56 json_config -- json_config/json_config.sh@385 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:06.515 08:43:56 json_config -- json_config/json_config.sh@385 -- # tgt_rpc save_config 00:08:06.515 08:43:56 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:08:06.515 + '[' 2 -ne 2 ']' 00:08:06.515 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:08:06.515 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:08:06.515 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:08:06.515 +++ basename /dev/fd/62 00:08:06.515 ++ mktemp /tmp/62.XXX 00:08:06.515 + tmp_file_1=/tmp/62.Hwg 00:08:06.515 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:06.515 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:08:06.515 + tmp_file_2=/tmp/spdk_tgt_config.json.D1c 00:08:06.515 + ret=0 00:08:06.515 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:08:06.776 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:08:07.036 + diff -u /tmp/62.Hwg /tmp/spdk_tgt_config.json.D1c 00:08:07.036 + echo 'INFO: JSON config files are the same' 00:08:07.036 INFO: JSON config files are the same 00:08:07.036 + rm /tmp/62.Hwg /tmp/spdk_tgt_config.json.D1c 00:08:07.036 + exit 0 00:08:07.036 08:43:56 json_config -- json_config/json_config.sh@386 -- # [[ 0 -eq 1 ]] 00:08:07.036 08:43:56 json_config -- json_config/json_config.sh@391 -- # echo 'INFO: changing configuration and checking if this can be detected...' 00:08:07.036 INFO: changing configuration and checking if this can be detected... 00:08:07.036 08:43:56 json_config -- json_config/json_config.sh@393 -- # tgt_rpc bdev_malloc_delete MallocBdevForConfigChangeCheck 00:08:07.036 08:43:56 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_delete MallocBdevForConfigChangeCheck 00:08:07.036 08:43:57 json_config -- json_config/json_config.sh@394 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:07.037 08:43:57 json_config -- json_config/json_config.sh@394 -- # tgt_rpc save_config 00:08:07.037 08:43:57 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:08:07.037 + '[' 2 -ne 2 ']' 00:08:07.037 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:08:07.037 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:08:07.037 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:08:07.037 +++ basename /dev/fd/62 00:08:07.037 ++ mktemp /tmp/62.XXX 00:08:07.037 + tmp_file_1=/tmp/62.bwd 00:08:07.037 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:07.037 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:08:07.037 + tmp_file_2=/tmp/spdk_tgt_config.json.Uqj 00:08:07.037 + ret=0 00:08:07.037 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:08:07.607 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:08:07.607 + diff -u /tmp/62.bwd /tmp/spdk_tgt_config.json.Uqj 00:08:07.607 + ret=1 00:08:07.607 + echo '=== Start of file: /tmp/62.bwd ===' 00:08:07.607 + cat /tmp/62.bwd 00:08:07.607 + echo '=== End of file: /tmp/62.bwd ===' 00:08:07.607 + echo '' 00:08:07.607 + echo '=== Start of file: /tmp/spdk_tgt_config.json.Uqj ===' 00:08:07.607 + cat /tmp/spdk_tgt_config.json.Uqj 00:08:07.607 + echo '=== End of file: /tmp/spdk_tgt_config.json.Uqj ===' 00:08:07.607 + echo '' 00:08:07.607 + rm /tmp/62.bwd /tmp/spdk_tgt_config.json.Uqj 00:08:07.607 + exit 1 00:08:07.607 08:43:57 json_config -- json_config/json_config.sh@398 -- # echo 'INFO: configuration change detected.' 00:08:07.607 INFO: configuration change detected. 00:08:07.607 08:43:57 json_config -- json_config/json_config.sh@401 -- # json_config_test_fini 00:08:07.607 08:43:57 json_config -- json_config/json_config.sh@313 -- # timing_enter json_config_test_fini 00:08:07.607 08:43:57 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:08:07.607 08:43:57 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:07.607 08:43:57 json_config -- json_config/json_config.sh@314 -- # local ret=0 00:08:07.607 08:43:57 json_config -- json_config/json_config.sh@316 -- # [[ -n '' ]] 00:08:07.607 08:43:57 json_config -- json_config/json_config.sh@324 -- # [[ -n 1658091 ]] 00:08:07.607 08:43:57 json_config -- json_config/json_config.sh@327 -- # cleanup_bdev_subsystem_config 00:08:07.607 08:43:57 json_config -- json_config/json_config.sh@191 -- # timing_enter cleanup_bdev_subsystem_config 00:08:07.607 08:43:57 json_config -- common/autotest_common.sh@724 -- # xtrace_disable 00:08:07.607 08:43:57 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:07.607 08:43:57 json_config -- json_config/json_config.sh@193 -- # [[ 0 -eq 1 ]] 00:08:07.607 08:43:57 json_config -- json_config/json_config.sh@200 -- # uname -s 00:08:07.608 08:43:57 json_config -- json_config/json_config.sh@200 -- # [[ Linux = Linux ]] 00:08:07.608 08:43:57 json_config -- json_config/json_config.sh@201 -- # rm -f /sample_aio 00:08:07.608 08:43:57 json_config -- json_config/json_config.sh@204 -- # [[ 0 -eq 1 ]] 00:08:07.608 08:43:57 json_config -- json_config/json_config.sh@208 -- # timing_exit cleanup_bdev_subsystem_config 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:07.608 08:43:57 json_config -- json_config/json_config.sh@330 -- # killprocess 1658091 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@950 -- # '[' -z 1658091 ']' 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@954 -- # kill -0 1658091 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@955 -- # uname 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1658091 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1658091' 00:08:07.608 killing process with pid 1658091 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@969 -- # kill 1658091 00:08:07.608 08:43:57 json_config -- common/autotest_common.sh@974 -- # wait 1658091 00:08:07.868 08:43:57 json_config -- json_config/json_config.sh@333 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:08:07.868 08:43:57 json_config -- json_config/json_config.sh@334 -- # timing_exit json_config_test_fini 00:08:07.868 08:43:57 json_config -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:07.868 08:43:57 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:07.868 08:43:57 json_config -- json_config/json_config.sh@335 -- # return 0 00:08:07.868 08:43:57 json_config -- json_config/json_config.sh@403 -- # echo 'INFO: Success' 00:08:07.868 INFO: Success 00:08:07.868 00:08:07.868 real 0m7.443s 00:08:07.868 user 0m9.009s 00:08:07.868 sys 0m2.005s 00:08:07.868 08:43:57 json_config -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:07.868 08:43:57 json_config -- common/autotest_common.sh@10 -- # set +x 00:08:07.868 ************************************ 00:08:07.868 END TEST json_config 00:08:07.869 ************************************ 00:08:08.130 08:43:57 -- spdk/autotest.sh@160 -- # run_test json_config_extra_key /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:08:08.130 08:43:57 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:08.130 08:43:57 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:08.130 08:43:57 -- common/autotest_common.sh@10 -- # set +x 00:08:08.130 ************************************ 00:08:08.130 START TEST json_config_extra_key 00:08:08.130 ************************************ 00:08:08.130 08:43:58 json_config_extra_key -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:08:08.130 08:43:58 json_config_extra_key -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:08.130 08:43:58 json_config_extra_key -- common/autotest_common.sh@1689 -- # lcov --version 00:08:08.130 08:43:58 json_config_extra_key -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:08.130 08:43:58 json_config_extra_key -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@336 -- # IFS=.-: 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@336 -- # read -ra ver1 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@337 -- # IFS=.-: 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@337 -- # read -ra ver2 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@338 -- # local 'op=<' 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@340 -- # ver1_l=2 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@341 -- # ver2_l=1 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@344 -- # case "$op" in 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@345 -- # : 1 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@365 -- # decimal 1 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@353 -- # local d=1 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@355 -- # echo 1 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@365 -- # ver1[v]=1 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@366 -- # decimal 2 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@353 -- # local d=2 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@355 -- # echo 2 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@366 -- # ver2[v]=2 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@368 -- # return 0 00:08:08.130 08:43:58 json_config_extra_key -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:08.130 08:43:58 json_config_extra_key -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:08.130 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:08.130 --rc genhtml_branch_coverage=1 00:08:08.130 --rc genhtml_function_coverage=1 00:08:08.130 --rc genhtml_legend=1 00:08:08.130 --rc geninfo_all_blocks=1 00:08:08.130 --rc geninfo_unexecuted_blocks=1 00:08:08.130 00:08:08.130 ' 00:08:08.130 08:43:58 json_config_extra_key -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:08.130 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:08.130 --rc genhtml_branch_coverage=1 00:08:08.130 --rc genhtml_function_coverage=1 00:08:08.130 --rc genhtml_legend=1 00:08:08.130 --rc geninfo_all_blocks=1 00:08:08.130 --rc geninfo_unexecuted_blocks=1 00:08:08.130 00:08:08.130 ' 00:08:08.130 08:43:58 json_config_extra_key -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:08.130 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:08.130 --rc genhtml_branch_coverage=1 00:08:08.130 --rc genhtml_function_coverage=1 00:08:08.130 --rc genhtml_legend=1 00:08:08.130 --rc geninfo_all_blocks=1 00:08:08.130 --rc geninfo_unexecuted_blocks=1 00:08:08.130 00:08:08.130 ' 00:08:08.130 08:43:58 json_config_extra_key -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:08.130 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:08.130 --rc genhtml_branch_coverage=1 00:08:08.130 --rc genhtml_function_coverage=1 00:08:08.130 --rc genhtml_legend=1 00:08:08.130 --rc geninfo_all_blocks=1 00:08:08.130 --rc geninfo_unexecuted_blocks=1 00:08:08.130 00:08:08.130 ' 00:08:08.130 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:08.130 08:43:58 json_config_extra_key -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@15 -- # shopt -s extglob 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:08.130 08:43:58 json_config_extra_key -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:08.130 08:43:58 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:08.130 08:43:58 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:08.131 08:43:58 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:08.131 08:43:58 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:08:08.131 08:43:58 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:08.131 08:43:58 json_config_extra_key -- nvmf/common.sh@51 -- # : 0 00:08:08.131 08:43:58 json_config_extra_key -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:08:08.131 08:43:58 json_config_extra_key -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:08:08.131 08:43:58 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:08.131 08:43:58 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:08.131 08:43:58 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:08.131 08:43:58 json_config_extra_key -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:08:08.131 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:08:08.131 08:43:58 json_config_extra_key -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:08:08.131 08:43:58 json_config_extra_key -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:08:08.131 08:43:58 json_config_extra_key -- nvmf/common.sh@55 -- # have_pci_nics=0 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json') 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:08:08.131 INFO: launching applications... 00:08:08.131 08:43:58 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=1658573 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:08:08.131 Waiting for target to run... 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 1658573 /var/tmp/spdk_tgt.sock 00:08:08.131 08:43:58 json_config_extra_key -- common/autotest_common.sh@831 -- # '[' -z 1658573 ']' 00:08:08.131 08:43:58 json_config_extra_key -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:08:08.131 08:43:58 json_config_extra_key -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:08.131 08:43:58 json_config_extra_key -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:08:08.131 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:08:08.131 08:43:58 json_config_extra_key -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:08:08.131 08:43:58 json_config_extra_key -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:08.131 08:43:58 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:08:08.392 [2024-11-06 08:43:58.292925] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:08.392 [2024-11-06 08:43:58.292998] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1658573 ] 00:08:08.652 [2024-11-06 08:43:58.714271] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:08.652 [2024-11-06 08:43:58.750729] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:09.225 08:43:59 json_config_extra_key -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:09.225 08:43:59 json_config_extra_key -- common/autotest_common.sh@864 -- # return 0 00:08:09.225 08:43:59 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:08:09.225 00:08:09.225 08:43:59 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:08:09.225 INFO: shutting down applications... 00:08:09.225 08:43:59 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:08:09.225 08:43:59 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:08:09.225 08:43:59 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:08:09.225 08:43:59 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 1658573 ]] 00:08:09.225 08:43:59 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 1658573 00:08:09.225 08:43:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:08:09.225 08:43:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:08:09.225 08:43:59 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 1658573 00:08:09.225 08:43:59 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:08:09.796 08:43:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:08:09.796 08:43:59 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:08:09.796 08:43:59 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 1658573 00:08:09.796 08:43:59 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:08:09.796 08:43:59 json_config_extra_key -- json_config/common.sh@43 -- # break 00:08:09.796 08:43:59 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:08:09.796 08:43:59 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:08:09.796 SPDK target shutdown done 00:08:09.796 08:43:59 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:08:09.796 Success 00:08:09.796 00:08:09.796 real 0m1.592s 00:08:09.796 user 0m1.116s 00:08:09.796 sys 0m0.544s 00:08:09.796 08:43:59 json_config_extra_key -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:09.796 08:43:59 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:08:09.796 ************************************ 00:08:09.796 END TEST json_config_extra_key 00:08:09.796 ************************************ 00:08:09.796 08:43:59 -- spdk/autotest.sh@161 -- # run_test alias_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:08:09.796 08:43:59 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:09.796 08:43:59 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:09.796 08:43:59 -- common/autotest_common.sh@10 -- # set +x 00:08:09.796 ************************************ 00:08:09.796 START TEST alias_rpc 00:08:09.796 ************************************ 00:08:09.796 08:43:59 alias_rpc -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:08:09.796 * Looking for test storage... 00:08:09.796 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc 00:08:09.796 08:43:59 alias_rpc -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:09.796 08:43:59 alias_rpc -- common/autotest_common.sh@1689 -- # lcov --version 00:08:09.796 08:43:59 alias_rpc -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:09.796 08:43:59 alias_rpc -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:09.796 08:43:59 alias_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:09.796 08:43:59 alias_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:09.796 08:43:59 alias_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:09.796 08:43:59 alias_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:08:09.796 08:43:59 alias_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:08:09.796 08:43:59 alias_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:08:09.796 08:43:59 alias_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@344 -- # case "$op" in 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@345 -- # : 1 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@365 -- # decimal 1 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@353 -- # local d=1 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@355 -- # echo 1 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@366 -- # decimal 2 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@353 -- # local d=2 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@355 -- # echo 2 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:09.797 08:43:59 alias_rpc -- scripts/common.sh@368 -- # return 0 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:09.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:09.797 --rc genhtml_branch_coverage=1 00:08:09.797 --rc genhtml_function_coverage=1 00:08:09.797 --rc genhtml_legend=1 00:08:09.797 --rc geninfo_all_blocks=1 00:08:09.797 --rc geninfo_unexecuted_blocks=1 00:08:09.797 00:08:09.797 ' 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:09.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:09.797 --rc genhtml_branch_coverage=1 00:08:09.797 --rc genhtml_function_coverage=1 00:08:09.797 --rc genhtml_legend=1 00:08:09.797 --rc geninfo_all_blocks=1 00:08:09.797 --rc geninfo_unexecuted_blocks=1 00:08:09.797 00:08:09.797 ' 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:09.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:09.797 --rc genhtml_branch_coverage=1 00:08:09.797 --rc genhtml_function_coverage=1 00:08:09.797 --rc genhtml_legend=1 00:08:09.797 --rc geninfo_all_blocks=1 00:08:09.797 --rc geninfo_unexecuted_blocks=1 00:08:09.797 00:08:09.797 ' 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:09.797 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:09.797 --rc genhtml_branch_coverage=1 00:08:09.797 --rc genhtml_function_coverage=1 00:08:09.797 --rc genhtml_legend=1 00:08:09.797 --rc geninfo_all_blocks=1 00:08:09.797 --rc geninfo_unexecuted_blocks=1 00:08:09.797 00:08:09.797 ' 00:08:09.797 08:43:59 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:08:09.797 08:43:59 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=1658952 00:08:09.797 08:43:59 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 1658952 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@831 -- # '[' -z 1658952 ']' 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:09.797 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:09.797 08:43:59 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:09.797 08:43:59 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:08:10.057 [2024-11-06 08:43:59.933079] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:10.058 [2024-11-06 08:43:59.933155] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1658952 ] 00:08:10.058 [2024-11-06 08:44:00.009543] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:10.058 [2024-11-06 08:44:00.057861] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:10.629 08:44:00 alias_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:10.629 08:44:00 alias_rpc -- common/autotest_common.sh@864 -- # return 0 00:08:10.629 08:44:00 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_config -i 00:08:10.890 08:44:00 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 1658952 00:08:10.890 08:44:00 alias_rpc -- common/autotest_common.sh@950 -- # '[' -z 1658952 ']' 00:08:10.890 08:44:00 alias_rpc -- common/autotest_common.sh@954 -- # kill -0 1658952 00:08:10.890 08:44:00 alias_rpc -- common/autotest_common.sh@955 -- # uname 00:08:10.890 08:44:00 alias_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:10.890 08:44:00 alias_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1658952 00:08:10.890 08:44:00 alias_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:10.890 08:44:00 alias_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:10.890 08:44:00 alias_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1658952' 00:08:10.890 killing process with pid 1658952 00:08:10.890 08:44:00 alias_rpc -- common/autotest_common.sh@969 -- # kill 1658952 00:08:10.890 08:44:00 alias_rpc -- common/autotest_common.sh@974 -- # wait 1658952 00:08:11.151 00:08:11.151 real 0m1.480s 00:08:11.151 user 0m1.621s 00:08:11.151 sys 0m0.398s 00:08:11.151 08:44:01 alias_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:11.151 08:44:01 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:11.151 ************************************ 00:08:11.151 END TEST alias_rpc 00:08:11.151 ************************************ 00:08:11.151 08:44:01 -- spdk/autotest.sh@163 -- # [[ 0 -eq 0 ]] 00:08:11.151 08:44:01 -- spdk/autotest.sh@164 -- # run_test spdkcli_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:08:11.151 08:44:01 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:11.151 08:44:01 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:11.151 08:44:01 -- common/autotest_common.sh@10 -- # set +x 00:08:11.151 ************************************ 00:08:11.151 START TEST spdkcli_tcp 00:08:11.151 ************************************ 00:08:11.151 08:44:01 spdkcli_tcp -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:08:11.412 * Looking for test storage... 00:08:11.412 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@1689 -- # lcov --version 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@344 -- # case "$op" in 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@345 -- # : 1 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@365 -- # decimal 1 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@353 -- # local d=1 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@355 -- # echo 1 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@366 -- # decimal 2 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@353 -- # local d=2 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@355 -- # echo 2 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:11.412 08:44:01 spdkcli_tcp -- scripts/common.sh@368 -- # return 0 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:11.412 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:11.412 --rc genhtml_branch_coverage=1 00:08:11.412 --rc genhtml_function_coverage=1 00:08:11.412 --rc genhtml_legend=1 00:08:11.412 --rc geninfo_all_blocks=1 00:08:11.412 --rc geninfo_unexecuted_blocks=1 00:08:11.412 00:08:11.412 ' 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:11.412 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:11.412 --rc genhtml_branch_coverage=1 00:08:11.412 --rc genhtml_function_coverage=1 00:08:11.412 --rc genhtml_legend=1 00:08:11.412 --rc geninfo_all_blocks=1 00:08:11.412 --rc geninfo_unexecuted_blocks=1 00:08:11.412 00:08:11.412 ' 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:11.412 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:11.412 --rc genhtml_branch_coverage=1 00:08:11.412 --rc genhtml_function_coverage=1 00:08:11.412 --rc genhtml_legend=1 00:08:11.412 --rc geninfo_all_blocks=1 00:08:11.412 --rc geninfo_unexecuted_blocks=1 00:08:11.412 00:08:11.412 ' 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:11.412 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:11.412 --rc genhtml_branch_coverage=1 00:08:11.412 --rc genhtml_function_coverage=1 00:08:11.412 --rc genhtml_legend=1 00:08:11.412 --rc geninfo_all_blocks=1 00:08:11.412 --rc geninfo_unexecuted_blocks=1 00:08:11.412 00:08:11.412 ' 00:08:11.412 08:44:01 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:08:11.412 08:44:01 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:08:11.412 08:44:01 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:08:11.412 08:44:01 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:08:11.412 08:44:01 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:08:11.412 08:44:01 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:08:11.412 08:44:01 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:08:11.412 08:44:01 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=1659350 00:08:11.412 08:44:01 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 1659350 00:08:11.412 08:44:01 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:08:11.412 08:44:01 spdkcli_tcp -- common/autotest_common.sh@831 -- # '[' -z 1659350 ']' 00:08:11.413 08:44:01 spdkcli_tcp -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:11.413 08:44:01 spdkcli_tcp -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:11.413 08:44:01 spdkcli_tcp -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:11.413 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:11.413 08:44:01 spdkcli_tcp -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:11.413 08:44:01 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:08:11.413 [2024-11-06 08:44:01.514523] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:11.413 [2024-11-06 08:44:01.514590] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1659350 ] 00:08:11.673 [2024-11-06 08:44:01.589743] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:08:11.673 [2024-11-06 08:44:01.632795] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:11.673 [2024-11-06 08:44:01.632809] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:12.245 08:44:02 spdkcli_tcp -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:12.245 08:44:02 spdkcli_tcp -- common/autotest_common.sh@864 -- # return 0 00:08:12.245 08:44:02 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=1659685 00:08:12.245 08:44:02 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:08:12.245 08:44:02 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:08:12.506 [ 00:08:12.506 "bdev_malloc_delete", 00:08:12.506 "bdev_malloc_create", 00:08:12.506 "bdev_null_resize", 00:08:12.506 "bdev_null_delete", 00:08:12.506 "bdev_null_create", 00:08:12.506 "bdev_nvme_cuse_unregister", 00:08:12.506 "bdev_nvme_cuse_register", 00:08:12.506 "bdev_opal_new_user", 00:08:12.506 "bdev_opal_set_lock_state", 00:08:12.506 "bdev_opal_delete", 00:08:12.506 "bdev_opal_get_info", 00:08:12.506 "bdev_opal_create", 00:08:12.506 "bdev_nvme_opal_revert", 00:08:12.506 "bdev_nvme_opal_init", 00:08:12.506 "bdev_nvme_send_cmd", 00:08:12.506 "bdev_nvme_set_keys", 00:08:12.506 "bdev_nvme_get_path_iostat", 00:08:12.506 "bdev_nvme_get_mdns_discovery_info", 00:08:12.506 "bdev_nvme_stop_mdns_discovery", 00:08:12.506 "bdev_nvme_start_mdns_discovery", 00:08:12.506 "bdev_nvme_set_multipath_policy", 00:08:12.506 "bdev_nvme_set_preferred_path", 00:08:12.506 "bdev_nvme_get_io_paths", 00:08:12.506 "bdev_nvme_remove_error_injection", 00:08:12.506 "bdev_nvme_add_error_injection", 00:08:12.506 "bdev_nvme_get_discovery_info", 00:08:12.506 "bdev_nvme_stop_discovery", 00:08:12.506 "bdev_nvme_start_discovery", 00:08:12.506 "bdev_nvme_get_controller_health_info", 00:08:12.506 "bdev_nvme_disable_controller", 00:08:12.506 "bdev_nvme_enable_controller", 00:08:12.506 "bdev_nvme_reset_controller", 00:08:12.506 "bdev_nvme_get_transport_statistics", 00:08:12.506 "bdev_nvme_apply_firmware", 00:08:12.506 "bdev_nvme_detach_controller", 00:08:12.506 "bdev_nvme_get_controllers", 00:08:12.506 "bdev_nvme_attach_controller", 00:08:12.506 "bdev_nvme_set_hotplug", 00:08:12.506 "bdev_nvme_set_options", 00:08:12.506 "bdev_passthru_delete", 00:08:12.506 "bdev_passthru_create", 00:08:12.506 "bdev_lvol_set_parent_bdev", 00:08:12.506 "bdev_lvol_set_parent", 00:08:12.506 "bdev_lvol_check_shallow_copy", 00:08:12.506 "bdev_lvol_start_shallow_copy", 00:08:12.506 "bdev_lvol_grow_lvstore", 00:08:12.506 "bdev_lvol_get_lvols", 00:08:12.506 "bdev_lvol_get_lvstores", 00:08:12.506 "bdev_lvol_delete", 00:08:12.506 "bdev_lvol_set_read_only", 00:08:12.506 "bdev_lvol_resize", 00:08:12.506 "bdev_lvol_decouple_parent", 00:08:12.506 "bdev_lvol_inflate", 00:08:12.506 "bdev_lvol_rename", 00:08:12.506 "bdev_lvol_clone_bdev", 00:08:12.506 "bdev_lvol_clone", 00:08:12.506 "bdev_lvol_snapshot", 00:08:12.506 "bdev_lvol_create", 00:08:12.506 "bdev_lvol_delete_lvstore", 00:08:12.506 "bdev_lvol_rename_lvstore", 00:08:12.506 "bdev_lvol_create_lvstore", 00:08:12.506 "bdev_raid_set_options", 00:08:12.506 "bdev_raid_remove_base_bdev", 00:08:12.506 "bdev_raid_add_base_bdev", 00:08:12.506 "bdev_raid_delete", 00:08:12.506 "bdev_raid_create", 00:08:12.506 "bdev_raid_get_bdevs", 00:08:12.506 "bdev_error_inject_error", 00:08:12.506 "bdev_error_delete", 00:08:12.506 "bdev_error_create", 00:08:12.506 "bdev_split_delete", 00:08:12.506 "bdev_split_create", 00:08:12.506 "bdev_delay_delete", 00:08:12.506 "bdev_delay_create", 00:08:12.506 "bdev_delay_update_latency", 00:08:12.506 "bdev_zone_block_delete", 00:08:12.506 "bdev_zone_block_create", 00:08:12.506 "blobfs_create", 00:08:12.506 "blobfs_detect", 00:08:12.506 "blobfs_set_cache_size", 00:08:12.506 "bdev_aio_delete", 00:08:12.506 "bdev_aio_rescan", 00:08:12.506 "bdev_aio_create", 00:08:12.506 "bdev_ftl_set_property", 00:08:12.506 "bdev_ftl_get_properties", 00:08:12.506 "bdev_ftl_get_stats", 00:08:12.506 "bdev_ftl_unmap", 00:08:12.506 "bdev_ftl_unload", 00:08:12.506 "bdev_ftl_delete", 00:08:12.506 "bdev_ftl_load", 00:08:12.506 "bdev_ftl_create", 00:08:12.506 "bdev_virtio_attach_controller", 00:08:12.506 "bdev_virtio_scsi_get_devices", 00:08:12.506 "bdev_virtio_detach_controller", 00:08:12.506 "bdev_virtio_blk_set_hotplug", 00:08:12.506 "bdev_iscsi_delete", 00:08:12.506 "bdev_iscsi_create", 00:08:12.506 "bdev_iscsi_set_options", 00:08:12.506 "accel_error_inject_error", 00:08:12.506 "ioat_scan_accel_module", 00:08:12.506 "dsa_scan_accel_module", 00:08:12.506 "iaa_scan_accel_module", 00:08:12.506 "vfu_virtio_create_fs_endpoint", 00:08:12.506 "vfu_virtio_create_scsi_endpoint", 00:08:12.506 "vfu_virtio_scsi_remove_target", 00:08:12.506 "vfu_virtio_scsi_add_target", 00:08:12.506 "vfu_virtio_create_blk_endpoint", 00:08:12.506 "vfu_virtio_delete_endpoint", 00:08:12.506 "keyring_file_remove_key", 00:08:12.506 "keyring_file_add_key", 00:08:12.506 "keyring_linux_set_options", 00:08:12.506 "fsdev_aio_delete", 00:08:12.506 "fsdev_aio_create", 00:08:12.506 "iscsi_get_histogram", 00:08:12.506 "iscsi_enable_histogram", 00:08:12.506 "iscsi_set_options", 00:08:12.506 "iscsi_get_auth_groups", 00:08:12.506 "iscsi_auth_group_remove_secret", 00:08:12.506 "iscsi_auth_group_add_secret", 00:08:12.506 "iscsi_delete_auth_group", 00:08:12.506 "iscsi_create_auth_group", 00:08:12.506 "iscsi_set_discovery_auth", 00:08:12.506 "iscsi_get_options", 00:08:12.506 "iscsi_target_node_request_logout", 00:08:12.507 "iscsi_target_node_set_redirect", 00:08:12.507 "iscsi_target_node_set_auth", 00:08:12.507 "iscsi_target_node_add_lun", 00:08:12.507 "iscsi_get_stats", 00:08:12.507 "iscsi_get_connections", 00:08:12.507 "iscsi_portal_group_set_auth", 00:08:12.507 "iscsi_start_portal_group", 00:08:12.507 "iscsi_delete_portal_group", 00:08:12.507 "iscsi_create_portal_group", 00:08:12.507 "iscsi_get_portal_groups", 00:08:12.507 "iscsi_delete_target_node", 00:08:12.507 "iscsi_target_node_remove_pg_ig_maps", 00:08:12.507 "iscsi_target_node_add_pg_ig_maps", 00:08:12.507 "iscsi_create_target_node", 00:08:12.507 "iscsi_get_target_nodes", 00:08:12.507 "iscsi_delete_initiator_group", 00:08:12.507 "iscsi_initiator_group_remove_initiators", 00:08:12.507 "iscsi_initiator_group_add_initiators", 00:08:12.507 "iscsi_create_initiator_group", 00:08:12.507 "iscsi_get_initiator_groups", 00:08:12.507 "nvmf_set_crdt", 00:08:12.507 "nvmf_set_config", 00:08:12.507 "nvmf_set_max_subsystems", 00:08:12.507 "nvmf_stop_mdns_prr", 00:08:12.507 "nvmf_publish_mdns_prr", 00:08:12.507 "nvmf_subsystem_get_listeners", 00:08:12.507 "nvmf_subsystem_get_qpairs", 00:08:12.507 "nvmf_subsystem_get_controllers", 00:08:12.507 "nvmf_get_stats", 00:08:12.507 "nvmf_get_transports", 00:08:12.507 "nvmf_create_transport", 00:08:12.507 "nvmf_get_targets", 00:08:12.507 "nvmf_delete_target", 00:08:12.507 "nvmf_create_target", 00:08:12.507 "nvmf_subsystem_allow_any_host", 00:08:12.507 "nvmf_subsystem_set_keys", 00:08:12.507 "nvmf_subsystem_remove_host", 00:08:12.507 "nvmf_subsystem_add_host", 00:08:12.507 "nvmf_ns_remove_host", 00:08:12.507 "nvmf_ns_add_host", 00:08:12.507 "nvmf_subsystem_remove_ns", 00:08:12.507 "nvmf_subsystem_set_ns_ana_group", 00:08:12.507 "nvmf_subsystem_add_ns", 00:08:12.507 "nvmf_subsystem_listener_set_ana_state", 00:08:12.507 "nvmf_discovery_get_referrals", 00:08:12.507 "nvmf_discovery_remove_referral", 00:08:12.507 "nvmf_discovery_add_referral", 00:08:12.507 "nvmf_subsystem_remove_listener", 00:08:12.507 "nvmf_subsystem_add_listener", 00:08:12.507 "nvmf_delete_subsystem", 00:08:12.507 "nvmf_create_subsystem", 00:08:12.507 "nvmf_get_subsystems", 00:08:12.507 "env_dpdk_get_mem_stats", 00:08:12.507 "nbd_get_disks", 00:08:12.507 "nbd_stop_disk", 00:08:12.507 "nbd_start_disk", 00:08:12.507 "ublk_recover_disk", 00:08:12.507 "ublk_get_disks", 00:08:12.507 "ublk_stop_disk", 00:08:12.507 "ublk_start_disk", 00:08:12.507 "ublk_destroy_target", 00:08:12.507 "ublk_create_target", 00:08:12.507 "virtio_blk_create_transport", 00:08:12.507 "virtio_blk_get_transports", 00:08:12.507 "vhost_controller_set_coalescing", 00:08:12.507 "vhost_get_controllers", 00:08:12.507 "vhost_delete_controller", 00:08:12.507 "vhost_create_blk_controller", 00:08:12.507 "vhost_scsi_controller_remove_target", 00:08:12.507 "vhost_scsi_controller_add_target", 00:08:12.507 "vhost_start_scsi_controller", 00:08:12.507 "vhost_create_scsi_controller", 00:08:12.507 "thread_set_cpumask", 00:08:12.507 "scheduler_set_options", 00:08:12.507 "framework_get_governor", 00:08:12.507 "framework_get_scheduler", 00:08:12.507 "framework_set_scheduler", 00:08:12.507 "framework_get_reactors", 00:08:12.507 "thread_get_io_channels", 00:08:12.507 "thread_get_pollers", 00:08:12.507 "thread_get_stats", 00:08:12.507 "framework_monitor_context_switch", 00:08:12.507 "spdk_kill_instance", 00:08:12.507 "log_enable_timestamps", 00:08:12.507 "log_get_flags", 00:08:12.507 "log_clear_flag", 00:08:12.507 "log_set_flag", 00:08:12.507 "log_get_level", 00:08:12.507 "log_set_level", 00:08:12.507 "log_get_print_level", 00:08:12.507 "log_set_print_level", 00:08:12.507 "framework_enable_cpumask_locks", 00:08:12.507 "framework_disable_cpumask_locks", 00:08:12.507 "framework_wait_init", 00:08:12.507 "framework_start_init", 00:08:12.507 "scsi_get_devices", 00:08:12.507 "bdev_get_histogram", 00:08:12.507 "bdev_enable_histogram", 00:08:12.507 "bdev_set_qos_limit", 00:08:12.507 "bdev_set_qd_sampling_period", 00:08:12.507 "bdev_get_bdevs", 00:08:12.507 "bdev_reset_iostat", 00:08:12.507 "bdev_get_iostat", 00:08:12.507 "bdev_examine", 00:08:12.507 "bdev_wait_for_examine", 00:08:12.507 "bdev_set_options", 00:08:12.507 "accel_get_stats", 00:08:12.507 "accel_set_options", 00:08:12.507 "accel_set_driver", 00:08:12.507 "accel_crypto_key_destroy", 00:08:12.507 "accel_crypto_keys_get", 00:08:12.507 "accel_crypto_key_create", 00:08:12.507 "accel_assign_opc", 00:08:12.507 "accel_get_module_info", 00:08:12.507 "accel_get_opc_assignments", 00:08:12.507 "vmd_rescan", 00:08:12.507 "vmd_remove_device", 00:08:12.507 "vmd_enable", 00:08:12.507 "sock_get_default_impl", 00:08:12.507 "sock_set_default_impl", 00:08:12.507 "sock_impl_set_options", 00:08:12.507 "sock_impl_get_options", 00:08:12.507 "iobuf_get_stats", 00:08:12.507 "iobuf_set_options", 00:08:12.507 "keyring_get_keys", 00:08:12.507 "vfu_tgt_set_base_path", 00:08:12.507 "framework_get_pci_devices", 00:08:12.507 "framework_get_config", 00:08:12.507 "framework_get_subsystems", 00:08:12.507 "fsdev_set_opts", 00:08:12.507 "fsdev_get_opts", 00:08:12.507 "trace_get_info", 00:08:12.507 "trace_get_tpoint_group_mask", 00:08:12.507 "trace_disable_tpoint_group", 00:08:12.507 "trace_enable_tpoint_group", 00:08:12.507 "trace_clear_tpoint_mask", 00:08:12.507 "trace_set_tpoint_mask", 00:08:12.507 "notify_get_notifications", 00:08:12.507 "notify_get_types", 00:08:12.507 "spdk_get_version", 00:08:12.507 "rpc_get_methods" 00:08:12.507 ] 00:08:12.507 08:44:02 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:08:12.507 08:44:02 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:08:12.507 08:44:02 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 1659350 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@950 -- # '[' -z 1659350 ']' 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@954 -- # kill -0 1659350 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@955 -- # uname 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1659350 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1659350' 00:08:12.507 killing process with pid 1659350 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@969 -- # kill 1659350 00:08:12.507 08:44:02 spdkcli_tcp -- common/autotest_common.sh@974 -- # wait 1659350 00:08:12.767 00:08:12.767 real 0m1.519s 00:08:12.767 user 0m2.761s 00:08:12.767 sys 0m0.431s 00:08:12.767 08:44:02 spdkcli_tcp -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:12.767 08:44:02 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:08:12.767 ************************************ 00:08:12.767 END TEST spdkcli_tcp 00:08:12.767 ************************************ 00:08:12.767 08:44:02 -- spdk/autotest.sh@167 -- # run_test dpdk_mem_utility /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:08:12.768 08:44:02 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:12.768 08:44:02 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:12.768 08:44:02 -- common/autotest_common.sh@10 -- # set +x 00:08:12.768 ************************************ 00:08:12.768 START TEST dpdk_mem_utility 00:08:12.768 ************************************ 00:08:12.768 08:44:02 dpdk_mem_utility -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:08:13.028 * Looking for test storage... 00:08:13.028 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility 00:08:13.028 08:44:02 dpdk_mem_utility -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:13.028 08:44:02 dpdk_mem_utility -- common/autotest_common.sh@1689 -- # lcov --version 00:08:13.028 08:44:02 dpdk_mem_utility -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:13.028 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@336 -- # IFS=.-: 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@336 -- # read -ra ver1 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@337 -- # IFS=.-: 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@337 -- # read -ra ver2 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@338 -- # local 'op=<' 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@340 -- # ver1_l=2 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@341 -- # ver2_l=1 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@344 -- # case "$op" in 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@345 -- # : 1 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@365 -- # decimal 1 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=1 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 1 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@365 -- # ver1[v]=1 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@366 -- # decimal 2 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@353 -- # local d=2 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@355 -- # echo 2 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@366 -- # ver2[v]=2 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:13.028 08:44:03 dpdk_mem_utility -- scripts/common.sh@368 -- # return 0 00:08:13.028 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:13.028 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:13.028 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:13.028 --rc genhtml_branch_coverage=1 00:08:13.028 --rc genhtml_function_coverage=1 00:08:13.028 --rc genhtml_legend=1 00:08:13.028 --rc geninfo_all_blocks=1 00:08:13.028 --rc geninfo_unexecuted_blocks=1 00:08:13.028 00:08:13.028 ' 00:08:13.028 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:13.028 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:13.028 --rc genhtml_branch_coverage=1 00:08:13.029 --rc genhtml_function_coverage=1 00:08:13.029 --rc genhtml_legend=1 00:08:13.029 --rc geninfo_all_blocks=1 00:08:13.029 --rc geninfo_unexecuted_blocks=1 00:08:13.029 00:08:13.029 ' 00:08:13.029 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:13.029 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:13.029 --rc genhtml_branch_coverage=1 00:08:13.029 --rc genhtml_function_coverage=1 00:08:13.029 --rc genhtml_legend=1 00:08:13.029 --rc geninfo_all_blocks=1 00:08:13.029 --rc geninfo_unexecuted_blocks=1 00:08:13.029 00:08:13.029 ' 00:08:13.029 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:13.029 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:13.029 --rc genhtml_branch_coverage=1 00:08:13.029 --rc genhtml_function_coverage=1 00:08:13.029 --rc genhtml_legend=1 00:08:13.029 --rc geninfo_all_blocks=1 00:08:13.029 --rc geninfo_unexecuted_blocks=1 00:08:13.029 00:08:13.029 ' 00:08:13.029 08:44:03 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:08:13.029 08:44:03 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=1659763 00:08:13.029 08:44:03 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 1659763 00:08:13.029 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@831 -- # '[' -z 1659763 ']' 00:08:13.029 08:44:03 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:08:13.029 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:13.029 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:13.029 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:13.029 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:13.029 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:13.029 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:08:13.029 [2024-11-06 08:44:03.099832] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:13.029 [2024-11-06 08:44:03.099904] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1659763 ] 00:08:13.289 [2024-11-06 08:44:03.175487] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:13.289 [2024-11-06 08:44:03.217178] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:13.860 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:13.860 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@864 -- # return 0 00:08:13.860 08:44:03 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:08:13.860 08:44:03 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:08:13.860 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:13.860 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:08:13.860 { 00:08:13.860 "filename": "/tmp/spdk_mem_dump.txt" 00:08:13.860 } 00:08:13.860 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:13.860 08:44:03 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:08:13.860 DPDK memory size 818.000000 MiB in 1 heap(s) 00:08:13.860 1 heaps totaling size 818.000000 MiB 00:08:13.860 size: 818.000000 MiB heap id: 0 00:08:13.860 end heaps---------- 00:08:13.860 9 mempools totaling size 603.782043 MiB 00:08:13.860 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:08:13.860 size: 158.602051 MiB name: PDU_data_out_Pool 00:08:13.860 size: 100.555481 MiB name: bdev_io_1659763 00:08:13.860 size: 50.003479 MiB name: msgpool_1659763 00:08:13.860 size: 36.509338 MiB name: fsdev_io_1659763 00:08:13.860 size: 21.763794 MiB name: PDU_Pool 00:08:13.860 size: 19.513306 MiB name: SCSI_TASK_Pool 00:08:13.860 size: 4.133484 MiB name: evtpool_1659763 00:08:13.860 size: 0.026123 MiB name: Session_Pool 00:08:13.860 end mempools------- 00:08:13.860 6 memzones totaling size 4.142822 MiB 00:08:13.860 size: 1.000366 MiB name: RG_ring_0_1659763 00:08:13.860 size: 1.000366 MiB name: RG_ring_1_1659763 00:08:13.860 size: 1.000366 MiB name: RG_ring_4_1659763 00:08:13.860 size: 1.000366 MiB name: RG_ring_5_1659763 00:08:13.860 size: 0.125366 MiB name: RG_ring_2_1659763 00:08:13.860 size: 0.015991 MiB name: RG_ring_3_1659763 00:08:13.860 end memzones------- 00:08:13.860 08:44:03 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py -m 0 00:08:14.121 heap id: 0 total size: 818.000000 MiB number of busy elements: 44 number of free elements: 15 00:08:14.121 list of free elements. size: 10.852478 MiB 00:08:14.121 element at address: 0x200019200000 with size: 0.999878 MiB 00:08:14.121 element at address: 0x200019400000 with size: 0.999878 MiB 00:08:14.121 element at address: 0x200000400000 with size: 0.998535 MiB 00:08:14.121 element at address: 0x200032000000 with size: 0.994446 MiB 00:08:14.121 element at address: 0x200006400000 with size: 0.959839 MiB 00:08:14.121 element at address: 0x200012c00000 with size: 0.944275 MiB 00:08:14.121 element at address: 0x200019600000 with size: 0.936584 MiB 00:08:14.121 element at address: 0x200000200000 with size: 0.717346 MiB 00:08:14.121 element at address: 0x20001ae00000 with size: 0.582886 MiB 00:08:14.121 element at address: 0x200000c00000 with size: 0.495422 MiB 00:08:14.121 element at address: 0x20000a600000 with size: 0.490723 MiB 00:08:14.121 element at address: 0x200019800000 with size: 0.485657 MiB 00:08:14.121 element at address: 0x200003e00000 with size: 0.481934 MiB 00:08:14.121 element at address: 0x200028200000 with size: 0.410034 MiB 00:08:14.121 element at address: 0x200000800000 with size: 0.355042 MiB 00:08:14.121 list of standard malloc elements. size: 199.218628 MiB 00:08:14.121 element at address: 0x20000a7fff80 with size: 132.000122 MiB 00:08:14.121 element at address: 0x2000065fff80 with size: 64.000122 MiB 00:08:14.121 element at address: 0x2000192fff80 with size: 1.000122 MiB 00:08:14.121 element at address: 0x2000194fff80 with size: 1.000122 MiB 00:08:14.121 element at address: 0x2000196fff80 with size: 1.000122 MiB 00:08:14.121 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:08:14.121 element at address: 0x2000196eff00 with size: 0.062622 MiB 00:08:14.121 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:08:14.121 element at address: 0x2000196efdc0 with size: 0.000305 MiB 00:08:14.121 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000004ffa00 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000004ffac0 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000004ffb80 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000004ffd80 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000004ffe40 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20000085ae40 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20000085b040 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20000085f300 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20000087f5c0 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20000087f680 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000008ff940 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000008ffb40 with size: 0.000183 MiB 00:08:14.121 element at address: 0x200000c7ed40 with size: 0.000183 MiB 00:08:14.121 element at address: 0x200000cff000 with size: 0.000183 MiB 00:08:14.121 element at address: 0x200000cff0c0 with size: 0.000183 MiB 00:08:14.121 element at address: 0x200003e7b600 with size: 0.000183 MiB 00:08:14.121 element at address: 0x200003e7b6c0 with size: 0.000183 MiB 00:08:14.121 element at address: 0x200003efb980 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000064fdd80 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20000a67da00 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20000a67dac0 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20000a6fdd80 with size: 0.000183 MiB 00:08:14.121 element at address: 0x200012cf1bc0 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000196efc40 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000196efd00 with size: 0.000183 MiB 00:08:14.121 element at address: 0x2000198bc740 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20001ae95380 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20001ae95440 with size: 0.000183 MiB 00:08:14.121 element at address: 0x200028268f80 with size: 0.000183 MiB 00:08:14.121 element at address: 0x200028269040 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20002826fc40 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20002826fe40 with size: 0.000183 MiB 00:08:14.121 element at address: 0x20002826ff00 with size: 0.000183 MiB 00:08:14.121 list of memzone associated elements. size: 607.928894 MiB 00:08:14.121 element at address: 0x20001ae95500 with size: 211.416748 MiB 00:08:14.121 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:08:14.121 element at address: 0x20002826ffc0 with size: 157.562561 MiB 00:08:14.121 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:08:14.121 element at address: 0x200012df1e80 with size: 100.055054 MiB 00:08:14.121 associated memzone info: size: 100.054932 MiB name: MP_bdev_io_1659763_0 00:08:14.121 element at address: 0x200000dff380 with size: 48.003052 MiB 00:08:14.122 associated memzone info: size: 48.002930 MiB name: MP_msgpool_1659763_0 00:08:14.122 element at address: 0x200003ffdb80 with size: 36.008911 MiB 00:08:14.122 associated memzone info: size: 36.008789 MiB name: MP_fsdev_io_1659763_0 00:08:14.122 element at address: 0x2000199be940 with size: 20.255554 MiB 00:08:14.122 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:08:14.122 element at address: 0x2000321feb40 with size: 18.005066 MiB 00:08:14.122 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:08:14.122 element at address: 0x2000004fff00 with size: 3.000244 MiB 00:08:14.122 associated memzone info: size: 3.000122 MiB name: MP_evtpool_1659763_0 00:08:14.122 element at address: 0x2000009ffe00 with size: 2.000488 MiB 00:08:14.122 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_1659763 00:08:14.122 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:08:14.122 associated memzone info: size: 1.007996 MiB name: MP_evtpool_1659763 00:08:14.122 element at address: 0x20000a6fde40 with size: 1.008118 MiB 00:08:14.122 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:08:14.122 element at address: 0x2000198bc800 with size: 1.008118 MiB 00:08:14.122 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:08:14.122 element at address: 0x2000064fde40 with size: 1.008118 MiB 00:08:14.122 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:08:14.122 element at address: 0x200003efba40 with size: 1.008118 MiB 00:08:14.122 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:08:14.122 element at address: 0x200000cff180 with size: 1.000488 MiB 00:08:14.122 associated memzone info: size: 1.000366 MiB name: RG_ring_0_1659763 00:08:14.122 element at address: 0x2000008ffc00 with size: 1.000488 MiB 00:08:14.122 associated memzone info: size: 1.000366 MiB name: RG_ring_1_1659763 00:08:14.122 element at address: 0x200012cf1c80 with size: 1.000488 MiB 00:08:14.122 associated memzone info: size: 1.000366 MiB name: RG_ring_4_1659763 00:08:14.122 element at address: 0x2000320fe940 with size: 1.000488 MiB 00:08:14.122 associated memzone info: size: 1.000366 MiB name: RG_ring_5_1659763 00:08:14.122 element at address: 0x20000087f740 with size: 0.500488 MiB 00:08:14.122 associated memzone info: size: 0.500366 MiB name: RG_MP_fsdev_io_1659763 00:08:14.122 element at address: 0x200000c7ee00 with size: 0.500488 MiB 00:08:14.122 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_1659763 00:08:14.122 element at address: 0x20000a67db80 with size: 0.500488 MiB 00:08:14.122 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:08:14.122 element at address: 0x200003e7b780 with size: 0.500488 MiB 00:08:14.122 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:08:14.122 element at address: 0x20001987c540 with size: 0.250488 MiB 00:08:14.122 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:08:14.122 element at address: 0x2000002b7a40 with size: 0.125488 MiB 00:08:14.122 associated memzone info: size: 0.125366 MiB name: RG_MP_evtpool_1659763 00:08:14.122 element at address: 0x20000085f3c0 with size: 0.125488 MiB 00:08:14.122 associated memzone info: size: 0.125366 MiB name: RG_ring_2_1659763 00:08:14.122 element at address: 0x2000064f5b80 with size: 0.031738 MiB 00:08:14.122 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:08:14.122 element at address: 0x200028269100 with size: 0.023743 MiB 00:08:14.122 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:08:14.122 element at address: 0x20000085b100 with size: 0.016113 MiB 00:08:14.122 associated memzone info: size: 0.015991 MiB name: RG_ring_3_1659763 00:08:14.122 element at address: 0x20002826f240 with size: 0.002441 MiB 00:08:14.122 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:08:14.122 element at address: 0x2000004ffc40 with size: 0.000305 MiB 00:08:14.122 associated memzone info: size: 0.000183 MiB name: MP_msgpool_1659763 00:08:14.122 element at address: 0x2000008ffa00 with size: 0.000305 MiB 00:08:14.122 associated memzone info: size: 0.000183 MiB name: MP_fsdev_io_1659763 00:08:14.122 element at address: 0x20000085af00 with size: 0.000305 MiB 00:08:14.122 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_1659763 00:08:14.122 element at address: 0x20002826fd00 with size: 0.000305 MiB 00:08:14.122 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:08:14.122 08:44:03 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:08:14.122 08:44:03 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 1659763 00:08:14.122 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@950 -- # '[' -z 1659763 ']' 00:08:14.122 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@954 -- # kill -0 1659763 00:08:14.122 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@955 -- # uname 00:08:14.122 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:14.122 08:44:03 dpdk_mem_utility -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1659763 00:08:14.122 08:44:04 dpdk_mem_utility -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:14.122 08:44:04 dpdk_mem_utility -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:14.122 08:44:04 dpdk_mem_utility -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1659763' 00:08:14.122 killing process with pid 1659763 00:08:14.122 08:44:04 dpdk_mem_utility -- common/autotest_common.sh@969 -- # kill 1659763 00:08:14.122 08:44:04 dpdk_mem_utility -- common/autotest_common.sh@974 -- # wait 1659763 00:08:14.384 00:08:14.384 real 0m1.419s 00:08:14.384 user 0m1.499s 00:08:14.384 sys 0m0.403s 00:08:14.384 08:44:04 dpdk_mem_utility -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:14.384 08:44:04 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:08:14.384 ************************************ 00:08:14.384 END TEST dpdk_mem_utility 00:08:14.384 ************************************ 00:08:14.384 08:44:04 -- spdk/autotest.sh@168 -- # run_test event /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:08:14.384 08:44:04 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:14.384 08:44:04 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:14.384 08:44:04 -- common/autotest_common.sh@10 -- # set +x 00:08:14.384 ************************************ 00:08:14.384 START TEST event 00:08:14.384 ************************************ 00:08:14.384 08:44:04 event -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:08:14.384 * Looking for test storage... 00:08:14.384 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:08:14.384 08:44:04 event -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:14.384 08:44:04 event -- common/autotest_common.sh@1689 -- # lcov --version 00:08:14.384 08:44:04 event -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:14.645 08:44:04 event -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:14.645 08:44:04 event -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:14.645 08:44:04 event -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:14.645 08:44:04 event -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:14.645 08:44:04 event -- scripts/common.sh@336 -- # IFS=.-: 00:08:14.645 08:44:04 event -- scripts/common.sh@336 -- # read -ra ver1 00:08:14.645 08:44:04 event -- scripts/common.sh@337 -- # IFS=.-: 00:08:14.645 08:44:04 event -- scripts/common.sh@337 -- # read -ra ver2 00:08:14.645 08:44:04 event -- scripts/common.sh@338 -- # local 'op=<' 00:08:14.645 08:44:04 event -- scripts/common.sh@340 -- # ver1_l=2 00:08:14.645 08:44:04 event -- scripts/common.sh@341 -- # ver2_l=1 00:08:14.645 08:44:04 event -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:14.645 08:44:04 event -- scripts/common.sh@344 -- # case "$op" in 00:08:14.645 08:44:04 event -- scripts/common.sh@345 -- # : 1 00:08:14.645 08:44:04 event -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:14.645 08:44:04 event -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:14.645 08:44:04 event -- scripts/common.sh@365 -- # decimal 1 00:08:14.645 08:44:04 event -- scripts/common.sh@353 -- # local d=1 00:08:14.645 08:44:04 event -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:14.645 08:44:04 event -- scripts/common.sh@355 -- # echo 1 00:08:14.645 08:44:04 event -- scripts/common.sh@365 -- # ver1[v]=1 00:08:14.645 08:44:04 event -- scripts/common.sh@366 -- # decimal 2 00:08:14.645 08:44:04 event -- scripts/common.sh@353 -- # local d=2 00:08:14.645 08:44:04 event -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:14.645 08:44:04 event -- scripts/common.sh@355 -- # echo 2 00:08:14.645 08:44:04 event -- scripts/common.sh@366 -- # ver2[v]=2 00:08:14.645 08:44:04 event -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:14.645 08:44:04 event -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:14.645 08:44:04 event -- scripts/common.sh@368 -- # return 0 00:08:14.645 08:44:04 event -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:14.645 08:44:04 event -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:14.645 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:14.645 --rc genhtml_branch_coverage=1 00:08:14.645 --rc genhtml_function_coverage=1 00:08:14.645 --rc genhtml_legend=1 00:08:14.645 --rc geninfo_all_blocks=1 00:08:14.645 --rc geninfo_unexecuted_blocks=1 00:08:14.645 00:08:14.645 ' 00:08:14.645 08:44:04 event -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:14.645 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:14.645 --rc genhtml_branch_coverage=1 00:08:14.645 --rc genhtml_function_coverage=1 00:08:14.645 --rc genhtml_legend=1 00:08:14.645 --rc geninfo_all_blocks=1 00:08:14.645 --rc geninfo_unexecuted_blocks=1 00:08:14.645 00:08:14.645 ' 00:08:14.645 08:44:04 event -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:14.645 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:14.645 --rc genhtml_branch_coverage=1 00:08:14.645 --rc genhtml_function_coverage=1 00:08:14.645 --rc genhtml_legend=1 00:08:14.645 --rc geninfo_all_blocks=1 00:08:14.645 --rc geninfo_unexecuted_blocks=1 00:08:14.645 00:08:14.645 ' 00:08:14.645 08:44:04 event -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:14.645 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:14.645 --rc genhtml_branch_coverage=1 00:08:14.645 --rc genhtml_function_coverage=1 00:08:14.645 --rc genhtml_legend=1 00:08:14.645 --rc geninfo_all_blocks=1 00:08:14.645 --rc geninfo_unexecuted_blocks=1 00:08:14.645 00:08:14.645 ' 00:08:14.645 08:44:04 event -- event/event.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/nbd_common.sh 00:08:14.645 08:44:04 event -- bdev/nbd_common.sh@6 -- # set -e 00:08:14.645 08:44:04 event -- event/event.sh@45 -- # run_test event_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:08:14.645 08:44:04 event -- common/autotest_common.sh@1101 -- # '[' 6 -le 1 ']' 00:08:14.645 08:44:04 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:14.645 08:44:04 event -- common/autotest_common.sh@10 -- # set +x 00:08:14.645 ************************************ 00:08:14.645 START TEST event_perf 00:08:14.645 ************************************ 00:08:14.645 08:44:04 event.event_perf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:08:14.645 Running I/O for 1 seconds...[2024-11-06 08:44:04.596098] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:14.645 [2024-11-06 08:44:04.596201] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1660168 ] 00:08:14.645 [2024-11-06 08:44:04.671183] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:14.645 [2024-11-06 08:44:04.710186] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:14.645 [2024-11-06 08:44:04.710300] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:08:14.645 [2024-11-06 08:44:04.710459] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:14.645 Running I/O for 1 seconds...[2024-11-06 08:44:04.710459] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:08:16.030 00:08:16.030 lcore 0: 183862 00:08:16.030 lcore 1: 183864 00:08:16.030 lcore 2: 183862 00:08:16.030 lcore 3: 183866 00:08:16.030 done. 00:08:16.030 00:08:16.030 real 0m1.169s 00:08:16.030 user 0m4.097s 00:08:16.030 sys 0m0.068s 00:08:16.030 08:44:05 event.event_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:16.030 08:44:05 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:08:16.030 ************************************ 00:08:16.030 END TEST event_perf 00:08:16.030 ************************************ 00:08:16.030 08:44:05 event -- event/event.sh@46 -- # run_test event_reactor /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:08:16.030 08:44:05 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:08:16.030 08:44:05 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:16.030 08:44:05 event -- common/autotest_common.sh@10 -- # set +x 00:08:16.030 ************************************ 00:08:16.030 START TEST event_reactor 00:08:16.030 ************************************ 00:08:16.030 08:44:05 event.event_reactor -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:08:16.030 [2024-11-06 08:44:05.845719] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:16.030 [2024-11-06 08:44:05.845824] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1660519 ] 00:08:16.030 [2024-11-06 08:44:05.921924] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:16.030 [2024-11-06 08:44:05.959019] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:16.971 test_start 00:08:16.971 oneshot 00:08:16.971 tick 100 00:08:16.971 tick 100 00:08:16.971 tick 250 00:08:16.971 tick 100 00:08:16.971 tick 100 00:08:16.971 tick 250 00:08:16.971 tick 100 00:08:16.971 tick 500 00:08:16.971 tick 100 00:08:16.971 tick 100 00:08:16.971 tick 250 00:08:16.971 tick 100 00:08:16.971 tick 100 00:08:16.971 test_end 00:08:16.971 00:08:16.971 real 0m1.167s 00:08:16.971 user 0m1.100s 00:08:16.971 sys 0m0.063s 00:08:16.971 08:44:06 event.event_reactor -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:16.971 08:44:06 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:08:16.971 ************************************ 00:08:16.971 END TEST event_reactor 00:08:16.971 ************************************ 00:08:16.971 08:44:07 event -- event/event.sh@47 -- # run_test event_reactor_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:08:16.971 08:44:07 event -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:08:16.971 08:44:07 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:16.971 08:44:07 event -- common/autotest_common.sh@10 -- # set +x 00:08:16.971 ************************************ 00:08:16.971 START TEST event_reactor_perf 00:08:16.971 ************************************ 00:08:16.971 08:44:07 event.event_reactor_perf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:08:17.232 [2024-11-06 08:44:07.085975] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:17.232 [2024-11-06 08:44:07.086072] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1660727 ] 00:08:17.232 [2024-11-06 08:44:07.162873] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:17.232 [2024-11-06 08:44:07.201460] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:18.174 test_start 00:08:18.174 test_end 00:08:18.174 Performance: 370306 events per second 00:08:18.174 00:08:18.174 real 0m1.169s 00:08:18.174 user 0m1.098s 00:08:18.174 sys 0m0.067s 00:08:18.174 08:44:08 event.event_reactor_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:18.174 08:44:08 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:08:18.174 ************************************ 00:08:18.174 END TEST event_reactor_perf 00:08:18.174 ************************************ 00:08:18.174 08:44:08 event -- event/event.sh@49 -- # uname -s 00:08:18.174 08:44:08 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:08:18.174 08:44:08 event -- event/event.sh@50 -- # run_test event_scheduler /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:08:18.174 08:44:08 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:18.174 08:44:08 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:18.174 08:44:08 event -- common/autotest_common.sh@10 -- # set +x 00:08:18.435 ************************************ 00:08:18.435 START TEST event_scheduler 00:08:18.435 ************************************ 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:08:18.435 * Looking for test storage... 00:08:18.435 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@1689 -- # lcov --version 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@336 -- # IFS=.-: 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@336 -- # read -ra ver1 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@337 -- # IFS=.-: 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@337 -- # read -ra ver2 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@338 -- # local 'op=<' 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@340 -- # ver1_l=2 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@341 -- # ver2_l=1 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@344 -- # case "$op" in 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@345 -- # : 1 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@365 -- # decimal 1 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@353 -- # local d=1 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@355 -- # echo 1 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@365 -- # ver1[v]=1 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@366 -- # decimal 2 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@353 -- # local d=2 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@355 -- # echo 2 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@366 -- # ver2[v]=2 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:18.435 08:44:08 event.event_scheduler -- scripts/common.sh@368 -- # return 0 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:18.435 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.435 --rc genhtml_branch_coverage=1 00:08:18.435 --rc genhtml_function_coverage=1 00:08:18.435 --rc genhtml_legend=1 00:08:18.435 --rc geninfo_all_blocks=1 00:08:18.435 --rc geninfo_unexecuted_blocks=1 00:08:18.435 00:08:18.435 ' 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:18.435 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.435 --rc genhtml_branch_coverage=1 00:08:18.435 --rc genhtml_function_coverage=1 00:08:18.435 --rc genhtml_legend=1 00:08:18.435 --rc geninfo_all_blocks=1 00:08:18.435 --rc geninfo_unexecuted_blocks=1 00:08:18.435 00:08:18.435 ' 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:18.435 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.435 --rc genhtml_branch_coverage=1 00:08:18.435 --rc genhtml_function_coverage=1 00:08:18.435 --rc genhtml_legend=1 00:08:18.435 --rc geninfo_all_blocks=1 00:08:18.435 --rc geninfo_unexecuted_blocks=1 00:08:18.435 00:08:18.435 ' 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:18.435 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:18.435 --rc genhtml_branch_coverage=1 00:08:18.435 --rc genhtml_function_coverage=1 00:08:18.435 --rc genhtml_legend=1 00:08:18.435 --rc geninfo_all_blocks=1 00:08:18.435 --rc geninfo_unexecuted_blocks=1 00:08:18.435 00:08:18.435 ' 00:08:18.435 08:44:08 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:08:18.435 08:44:08 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=1660997 00:08:18.435 08:44:08 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:08:18.435 08:44:08 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 1660997 00:08:18.435 08:44:08 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@831 -- # '[' -z 1660997 ']' 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:18.435 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:18.435 08:44:08 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:08:18.697 [2024-11-06 08:44:08.557289] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:18.697 [2024-11-06 08:44:08.557363] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1660997 ] 00:08:18.697 [2024-11-06 08:44:08.622814] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:18.697 [2024-11-06 08:44:08.664520] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:18.697 [2024-11-06 08:44:08.664556] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:18.697 [2024-11-06 08:44:08.664675] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:08:18.697 [2024-11-06 08:44:08.664678] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@864 -- # return 0 00:08:18.697 08:44:08 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:08:18.697 [2024-11-06 08:44:08.713283] dpdk_governor.c: 173:_init: *ERROR*: App core mask contains some but not all of a set of SMT siblings 00:08:18.697 [2024-11-06 08:44:08.713299] scheduler_dynamic.c: 280:init: *NOTICE*: Unable to initialize dpdk governor 00:08:18.697 [2024-11-06 08:44:08.713307] scheduler_dynamic.c: 427:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:08:18.697 [2024-11-06 08:44:08.713311] scheduler_dynamic.c: 429:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:08:18.697 [2024-11-06 08:44:08.713315] scheduler_dynamic.c: 431:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.697 08:44:08 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:08:18.697 [2024-11-06 08:44:08.775101] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.697 08:44:08 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:18.697 08:44:08 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:08:18.958 ************************************ 00:08:18.958 START TEST scheduler_create_thread 00:08:18.958 ************************************ 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1125 -- # scheduler_create_thread 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:18.958 2 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:18.958 3 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:18.958 4 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:18.958 5 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:18.958 6 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:18.958 7 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:18.958 8 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:08:18.958 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.959 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:18.959 9 00:08:18.959 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:18.959 08:44:08 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:08:18.959 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:18.959 08:44:08 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:19.529 10 00:08:19.529 08:44:09 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:19.529 08:44:09 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:08:19.529 08:44:09 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:19.529 08:44:09 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:20.912 08:44:10 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:20.912 08:44:10 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:08:20.912 08:44:10 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:08:20.912 08:44:10 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:20.912 08:44:10 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:21.484 08:44:11 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:21.484 08:44:11 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:08:21.484 08:44:11 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:21.484 08:44:11 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:22.425 08:44:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:22.425 08:44:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:08:22.425 08:44:12 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:08:22.425 08:44:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:22.425 08:44:12 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:23.005 08:44:13 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:23.005 00:08:23.005 real 0m4.225s 00:08:23.005 user 0m0.027s 00:08:23.005 sys 0m0.004s 00:08:23.005 08:44:13 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:23.005 08:44:13 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:08:23.005 ************************************ 00:08:23.005 END TEST scheduler_create_thread 00:08:23.005 ************************************ 00:08:23.005 08:44:13 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:08:23.005 08:44:13 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 1660997 00:08:23.005 08:44:13 event.event_scheduler -- common/autotest_common.sh@950 -- # '[' -z 1660997 ']' 00:08:23.005 08:44:13 event.event_scheduler -- common/autotest_common.sh@954 -- # kill -0 1660997 00:08:23.005 08:44:13 event.event_scheduler -- common/autotest_common.sh@955 -- # uname 00:08:23.005 08:44:13 event.event_scheduler -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:23.005 08:44:13 event.event_scheduler -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1660997 00:08:23.265 08:44:13 event.event_scheduler -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:08:23.265 08:44:13 event.event_scheduler -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:08:23.265 08:44:13 event.event_scheduler -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1660997' 00:08:23.265 killing process with pid 1660997 00:08:23.265 08:44:13 event.event_scheduler -- common/autotest_common.sh@969 -- # kill 1660997 00:08:23.265 08:44:13 event.event_scheduler -- common/autotest_common.sh@974 -- # wait 1660997 00:08:23.265 [2024-11-06 08:44:13.316453] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:08:23.526 00:08:23.526 real 0m5.164s 00:08:23.526 user 0m10.242s 00:08:23.526 sys 0m0.372s 00:08:23.526 08:44:13 event.event_scheduler -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:23.526 08:44:13 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:08:23.526 ************************************ 00:08:23.526 END TEST event_scheduler 00:08:23.526 ************************************ 00:08:23.526 08:44:13 event -- event/event.sh@51 -- # modprobe -n nbd 00:08:23.526 08:44:13 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:08:23.526 08:44:13 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:23.526 08:44:13 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:23.526 08:44:13 event -- common/autotest_common.sh@10 -- # set +x 00:08:23.526 ************************************ 00:08:23.526 START TEST app_repeat 00:08:23.526 ************************************ 00:08:23.526 08:44:13 event.app_repeat -- common/autotest_common.sh@1125 -- # app_repeat_test 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@19 -- # repeat_pid=1662042 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:08:23.526 08:44:13 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 1662042' 00:08:23.527 Process app_repeat pid: 1662042 00:08:23.527 08:44:13 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:08:23.527 08:44:13 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:08:23.527 spdk_app_start Round 0 00:08:23.527 08:44:13 event.app_repeat -- event/event.sh@25 -- # waitforlisten 1662042 /var/tmp/spdk-nbd.sock 00:08:23.527 08:44:13 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 1662042 ']' 00:08:23.527 08:44:13 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:08:23.527 08:44:13 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:23.527 08:44:13 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:08:23.527 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:08:23.527 08:44:13 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:23.527 08:44:13 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:08:23.527 [2024-11-06 08:44:13.594302] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:23.527 [2024-11-06 08:44:13.594374] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1662042 ] 00:08:23.787 [2024-11-06 08:44:13.668496] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:08:23.787 [2024-11-06 08:44:13.707604] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:23.787 [2024-11-06 08:44:13.707606] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:23.787 08:44:13 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:23.787 08:44:13 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:08:23.787 08:44:13 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:24.048 Malloc0 00:08:24.048 08:44:13 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:24.048 Malloc1 00:08:24.048 08:44:14 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:24.048 08:44:14 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:08:24.308 /dev/nbd0 00:08:24.308 08:44:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:08:24.308 08:44:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:08:24.308 08:44:14 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:08:24.308 08:44:14 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:08:24.308 08:44:14 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:08:24.308 08:44:14 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:08:24.308 08:44:14 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:08:24.308 08:44:14 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:08:24.308 08:44:14 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:08:24.308 08:44:14 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:08:24.308 08:44:14 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:24.308 1+0 records in 00:08:24.308 1+0 records out 00:08:24.308 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000278502 s, 14.7 MB/s 00:08:24.308 08:44:14 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:24.309 08:44:14 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:08:24.309 08:44:14 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:24.309 08:44:14 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:08:24.309 08:44:14 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:08:24.309 08:44:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:24.309 08:44:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:24.309 08:44:14 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:08:24.570 /dev/nbd1 00:08:24.570 08:44:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:08:24.570 08:44:14 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:24.570 1+0 records in 00:08:24.570 1+0 records out 00:08:24.570 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000249182 s, 16.4 MB/s 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:08:24.570 08:44:14 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:08:24.570 08:44:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:24.570 08:44:14 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:24.570 08:44:14 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:24.570 08:44:14 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:24.571 08:44:14 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:24.831 08:44:14 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:08:24.832 { 00:08:24.832 "nbd_device": "/dev/nbd0", 00:08:24.832 "bdev_name": "Malloc0" 00:08:24.832 }, 00:08:24.832 { 00:08:24.832 "nbd_device": "/dev/nbd1", 00:08:24.832 "bdev_name": "Malloc1" 00:08:24.832 } 00:08:24.832 ]' 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:08:24.832 { 00:08:24.832 "nbd_device": "/dev/nbd0", 00:08:24.832 "bdev_name": "Malloc0" 00:08:24.832 }, 00:08:24.832 { 00:08:24.832 "nbd_device": "/dev/nbd1", 00:08:24.832 "bdev_name": "Malloc1" 00:08:24.832 } 00:08:24.832 ]' 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:08:24.832 /dev/nbd1' 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:08:24.832 /dev/nbd1' 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:08:24.832 256+0 records in 00:08:24.832 256+0 records out 00:08:24.832 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0127715 s, 82.1 MB/s 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:08:24.832 256+0 records in 00:08:24.832 256+0 records out 00:08:24.832 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0165715 s, 63.3 MB/s 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:08:24.832 256+0 records in 00:08:24.832 256+0 records out 00:08:24.832 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0200379 s, 52.3 MB/s 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:24.832 08:44:14 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:08:25.093 08:44:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:08:25.093 08:44:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:08:25.093 08:44:15 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:08:25.093 08:44:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:25.093 08:44:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:25.093 08:44:15 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:08:25.093 08:44:15 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:08:25.093 08:44:15 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:08:25.093 08:44:15 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:25.093 08:44:15 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:08:25.353 08:44:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:25.614 08:44:15 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:08:25.614 08:44:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:08:25.614 08:44:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:25.614 08:44:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:08:25.614 08:44:15 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:08:25.614 08:44:15 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:08:25.614 08:44:15 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:08:25.614 08:44:15 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:08:25.614 08:44:15 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:08:25.614 08:44:15 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:08:25.614 08:44:15 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:08:25.874 [2024-11-06 08:44:15.773603] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:08:25.874 [2024-11-06 08:44:15.809118] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:25.874 [2024-11-06 08:44:15.809120] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:25.874 [2024-11-06 08:44:15.840702] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:08:25.874 [2024-11-06 08:44:15.840736] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:08:29.173 08:44:18 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:08:29.173 08:44:18 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:08:29.173 spdk_app_start Round 1 00:08:29.173 08:44:18 event.app_repeat -- event/event.sh@25 -- # waitforlisten 1662042 /var/tmp/spdk-nbd.sock 00:08:29.173 08:44:18 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 1662042 ']' 00:08:29.173 08:44:18 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:08:29.173 08:44:18 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:29.173 08:44:18 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:08:29.173 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:08:29.173 08:44:18 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:29.173 08:44:18 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:08:29.173 08:44:18 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:29.173 08:44:18 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:08:29.173 08:44:18 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:29.173 Malloc0 00:08:29.173 08:44:19 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:29.173 Malloc1 00:08:29.173 08:44:19 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:29.173 08:44:19 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:29.173 08:44:19 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:29.173 08:44:19 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:08:29.173 08:44:19 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:29.173 08:44:19 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:08:29.173 08:44:19 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:29.173 08:44:19 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:29.174 08:44:19 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:29.174 08:44:19 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:08:29.174 08:44:19 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:29.174 08:44:19 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:08:29.174 08:44:19 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:08:29.174 08:44:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:08:29.174 08:44:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:29.174 08:44:19 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:08:29.435 /dev/nbd0 00:08:29.435 08:44:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:08:29.435 08:44:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:29.435 1+0 records in 00:08:29.435 1+0 records out 00:08:29.435 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000100609 s, 40.7 MB/s 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:08:29.435 08:44:19 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:08:29.435 08:44:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:29.435 08:44:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:29.435 08:44:19 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:08:29.701 /dev/nbd1 00:08:29.702 08:44:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:08:29.702 08:44:19 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:29.702 1+0 records in 00:08:29.702 1+0 records out 00:08:29.702 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000160454 s, 25.5 MB/s 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:08:29.702 08:44:19 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:08:29.702 08:44:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:29.702 08:44:19 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:29.702 08:44:19 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:29.702 08:44:19 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:29.702 08:44:19 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:29.702 08:44:19 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:08:29.702 { 00:08:29.702 "nbd_device": "/dev/nbd0", 00:08:29.703 "bdev_name": "Malloc0" 00:08:29.703 }, 00:08:29.703 { 00:08:29.703 "nbd_device": "/dev/nbd1", 00:08:29.703 "bdev_name": "Malloc1" 00:08:29.703 } 00:08:29.703 ]' 00:08:29.703 08:44:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:08:29.703 { 00:08:29.703 "nbd_device": "/dev/nbd0", 00:08:29.703 "bdev_name": "Malloc0" 00:08:29.703 }, 00:08:29.703 { 00:08:29.703 "nbd_device": "/dev/nbd1", 00:08:29.703 "bdev_name": "Malloc1" 00:08:29.703 } 00:08:29.703 ]' 00:08:29.703 08:44:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:08:29.964 /dev/nbd1' 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:08:29.964 /dev/nbd1' 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:08:29.964 256+0 records in 00:08:29.964 256+0 records out 00:08:29.964 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0127315 s, 82.4 MB/s 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:08:29.964 256+0 records in 00:08:29.964 256+0 records out 00:08:29.964 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0158426 s, 66.2 MB/s 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:08:29.964 256+0 records in 00:08:29.964 256+0 records out 00:08:29.964 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0168821 s, 62.1 MB/s 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:29.964 08:44:19 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:30.225 08:44:20 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:08:30.484 08:44:20 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:08:30.484 08:44:20 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:08:30.744 08:44:20 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:08:30.744 [2024-11-06 08:44:20.819078] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:08:30.744 [2024-11-06 08:44:20.853622] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:30.744 [2024-11-06 08:44:20.853624] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:31.004 [2024-11-06 08:44:20.885989] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:08:31.004 [2024-11-06 08:44:20.886030] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:08:33.621 08:44:23 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:08:33.621 08:44:23 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:08:33.621 spdk_app_start Round 2 00:08:33.621 08:44:23 event.app_repeat -- event/event.sh@25 -- # waitforlisten 1662042 /var/tmp/spdk-nbd.sock 00:08:33.621 08:44:23 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 1662042 ']' 00:08:33.621 08:44:23 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:08:33.621 08:44:23 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:33.621 08:44:23 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:08:33.621 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:08:33.621 08:44:23 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:33.621 08:44:23 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:08:33.931 08:44:23 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:33.931 08:44:23 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:08:33.931 08:44:23 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:33.931 Malloc0 00:08:33.931 08:44:24 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:08:34.192 Malloc1 00:08:34.192 08:44:24 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:34.192 08:44:24 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:34.192 08:44:24 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:34.193 08:44:24 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:08:34.453 /dev/nbd0 00:08:34.453 08:44:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:08:34.453 08:44:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd0 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd0 /proc/partitions 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:34.453 1+0 records in 00:08:34.453 1+0 records out 00:08:34.453 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.00019837 s, 20.6 MB/s 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:08:34.453 08:44:24 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:08:34.453 08:44:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:34.453 08:44:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:34.453 08:44:24 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:08:34.714 /dev/nbd1 00:08:34.714 08:44:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:08:34.714 08:44:24 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@868 -- # local nbd_name=nbd1 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@869 -- # local i 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@871 -- # (( i = 1 )) 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@871 -- # (( i <= 20 )) 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@872 -- # grep -q -w nbd1 /proc/partitions 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@873 -- # break 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@884 -- # (( i = 1 )) 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@884 -- # (( i <= 20 )) 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@885 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:08:34.714 1+0 records in 00:08:34.714 1+0 records out 00:08:34.714 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000284538 s, 14.4 MB/s 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@886 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@886 -- # size=4096 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@887 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@888 -- # '[' 4096 '!=' 0 ']' 00:08:34.714 08:44:24 event.app_repeat -- common/autotest_common.sh@889 -- # return 0 00:08:34.714 08:44:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:08:34.714 08:44:24 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:08:34.714 08:44:24 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:34.714 08:44:24 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:34.714 08:44:24 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:34.714 08:44:24 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:08:34.714 { 00:08:34.714 "nbd_device": "/dev/nbd0", 00:08:34.714 "bdev_name": "Malloc0" 00:08:34.714 }, 00:08:34.714 { 00:08:34.714 "nbd_device": "/dev/nbd1", 00:08:34.714 "bdev_name": "Malloc1" 00:08:34.714 } 00:08:34.714 ]' 00:08:34.714 08:44:24 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:08:34.714 { 00:08:34.714 "nbd_device": "/dev/nbd0", 00:08:34.714 "bdev_name": "Malloc0" 00:08:34.714 }, 00:08:34.714 { 00:08:34.714 "nbd_device": "/dev/nbd1", 00:08:34.714 "bdev_name": "Malloc1" 00:08:34.714 } 00:08:34.714 ]' 00:08:34.714 08:44:24 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:08:34.974 /dev/nbd1' 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:08:34.974 /dev/nbd1' 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:34.974 08:44:24 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:08:34.975 256+0 records in 00:08:34.975 256+0 records out 00:08:34.975 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0127839 s, 82.0 MB/s 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:08:34.975 256+0 records in 00:08:34.975 256+0 records out 00:08:34.975 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0175456 s, 59.8 MB/s 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:08:34.975 256+0 records in 00:08:34.975 256+0 records out 00:08:34.975 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.019034 s, 55.1 MB/s 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:34.975 08:44:24 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:08:35.235 08:44:25 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:08:35.495 08:44:25 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:08:35.495 08:44:25 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:08:35.755 08:44:25 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:08:35.755 [2024-11-06 08:44:25.831924] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:08:35.755 [2024-11-06 08:44:25.866605] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:35.755 [2024-11-06 08:44:25.866607] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:36.015 [2024-11-06 08:44:25.898322] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:08:36.015 [2024-11-06 08:44:25.898356] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:08:39.321 08:44:28 event.app_repeat -- event/event.sh@38 -- # waitforlisten 1662042 /var/tmp/spdk-nbd.sock 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@831 -- # '[' -z 1662042 ']' 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:08:39.321 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@864 -- # return 0 00:08:39.321 08:44:28 event.app_repeat -- event/event.sh@39 -- # killprocess 1662042 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@950 -- # '[' -z 1662042 ']' 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@954 -- # kill -0 1662042 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@955 -- # uname 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1662042 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1662042' 00:08:39.321 killing process with pid 1662042 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@969 -- # kill 1662042 00:08:39.321 08:44:28 event.app_repeat -- common/autotest_common.sh@974 -- # wait 1662042 00:08:39.321 spdk_app_start is called in Round 0. 00:08:39.321 Shutdown signal received, stop current app iteration 00:08:39.321 Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 reinitialization... 00:08:39.321 spdk_app_start is called in Round 1. 00:08:39.321 Shutdown signal received, stop current app iteration 00:08:39.321 Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 reinitialization... 00:08:39.321 spdk_app_start is called in Round 2. 00:08:39.321 Shutdown signal received, stop current app iteration 00:08:39.321 Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 reinitialization... 00:08:39.321 spdk_app_start is called in Round 3. 00:08:39.321 Shutdown signal received, stop current app iteration 00:08:39.321 08:44:29 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:08:39.321 08:44:29 event.app_repeat -- event/event.sh@42 -- # return 0 00:08:39.321 00:08:39.321 real 0m15.498s 00:08:39.321 user 0m33.813s 00:08:39.321 sys 0m2.180s 00:08:39.321 08:44:29 event.app_repeat -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:39.321 08:44:29 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:08:39.321 ************************************ 00:08:39.321 END TEST app_repeat 00:08:39.321 ************************************ 00:08:39.321 08:44:29 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:08:39.321 08:44:29 event -- event/event.sh@55 -- # run_test cpu_locks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:08:39.321 08:44:29 event -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:39.321 08:44:29 event -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:39.321 08:44:29 event -- common/autotest_common.sh@10 -- # set +x 00:08:39.321 ************************************ 00:08:39.321 START TEST cpu_locks 00:08:39.321 ************************************ 00:08:39.321 08:44:29 event.cpu_locks -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:08:39.321 * Looking for test storage... 00:08:39.321 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:08:39.321 08:44:29 event.cpu_locks -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:39.321 08:44:29 event.cpu_locks -- common/autotest_common.sh@1689 -- # lcov --version 00:08:39.321 08:44:29 event.cpu_locks -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:39.321 08:44:29 event.cpu_locks -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@336 -- # IFS=.-: 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@336 -- # read -ra ver1 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@337 -- # IFS=.-: 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@337 -- # read -ra ver2 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@338 -- # local 'op=<' 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@340 -- # ver1_l=2 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@341 -- # ver2_l=1 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@344 -- # case "$op" in 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@345 -- # : 1 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@365 -- # decimal 1 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@353 -- # local d=1 00:08:39.321 08:44:29 event.cpu_locks -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:39.322 08:44:29 event.cpu_locks -- scripts/common.sh@355 -- # echo 1 00:08:39.322 08:44:29 event.cpu_locks -- scripts/common.sh@365 -- # ver1[v]=1 00:08:39.322 08:44:29 event.cpu_locks -- scripts/common.sh@366 -- # decimal 2 00:08:39.322 08:44:29 event.cpu_locks -- scripts/common.sh@353 -- # local d=2 00:08:39.322 08:44:29 event.cpu_locks -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:39.322 08:44:29 event.cpu_locks -- scripts/common.sh@355 -- # echo 2 00:08:39.322 08:44:29 event.cpu_locks -- scripts/common.sh@366 -- # ver2[v]=2 00:08:39.322 08:44:29 event.cpu_locks -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:39.322 08:44:29 event.cpu_locks -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:39.322 08:44:29 event.cpu_locks -- scripts/common.sh@368 -- # return 0 00:08:39.322 08:44:29 event.cpu_locks -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:39.322 08:44:29 event.cpu_locks -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:39.322 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:39.322 --rc genhtml_branch_coverage=1 00:08:39.322 --rc genhtml_function_coverage=1 00:08:39.322 --rc genhtml_legend=1 00:08:39.322 --rc geninfo_all_blocks=1 00:08:39.322 --rc geninfo_unexecuted_blocks=1 00:08:39.322 00:08:39.322 ' 00:08:39.322 08:44:29 event.cpu_locks -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:39.322 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:39.322 --rc genhtml_branch_coverage=1 00:08:39.322 --rc genhtml_function_coverage=1 00:08:39.322 --rc genhtml_legend=1 00:08:39.322 --rc geninfo_all_blocks=1 00:08:39.322 --rc geninfo_unexecuted_blocks=1 00:08:39.322 00:08:39.322 ' 00:08:39.322 08:44:29 event.cpu_locks -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:39.322 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:39.322 --rc genhtml_branch_coverage=1 00:08:39.322 --rc genhtml_function_coverage=1 00:08:39.322 --rc genhtml_legend=1 00:08:39.322 --rc geninfo_all_blocks=1 00:08:39.322 --rc geninfo_unexecuted_blocks=1 00:08:39.322 00:08:39.322 ' 00:08:39.322 08:44:29 event.cpu_locks -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:39.322 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:39.322 --rc genhtml_branch_coverage=1 00:08:39.322 --rc genhtml_function_coverage=1 00:08:39.322 --rc genhtml_legend=1 00:08:39.322 --rc geninfo_all_blocks=1 00:08:39.322 --rc geninfo_unexecuted_blocks=1 00:08:39.322 00:08:39.322 ' 00:08:39.322 08:44:29 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:08:39.322 08:44:29 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:08:39.322 08:44:29 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:08:39.322 08:44:29 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:08:39.322 08:44:29 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:39.322 08:44:29 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:39.322 08:44:29 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:39.322 ************************************ 00:08:39.322 START TEST default_locks 00:08:39.322 ************************************ 00:08:39.322 08:44:29 event.cpu_locks.default_locks -- common/autotest_common.sh@1125 -- # default_locks 00:08:39.322 08:44:29 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=1665591 00:08:39.322 08:44:29 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 1665591 00:08:39.322 08:44:29 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:08:39.322 08:44:29 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 1665591 ']' 00:08:39.322 08:44:29 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:39.322 08:44:29 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:39.322 08:44:29 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:39.322 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:39.322 08:44:29 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:39.322 08:44:29 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:08:39.322 [2024-11-06 08:44:29.413944] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:39.322 [2024-11-06 08:44:29.414007] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1665591 ] 00:08:39.582 [2024-11-06 08:44:29.487522] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:39.582 [2024-11-06 08:44:29.526410] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:40.152 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:40.152 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 0 00:08:40.152 08:44:30 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 1665591 00:08:40.152 08:44:30 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 1665591 00:08:40.152 08:44:30 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:40.414 lslocks: write error 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 1665591 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@950 -- # '[' -z 1665591 ']' 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # kill -0 1665591 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # uname 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1665591 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1665591' 00:08:40.414 killing process with pid 1665591 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@969 -- # kill 1665591 00:08:40.414 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@974 -- # wait 1665591 00:08:40.675 08:44:30 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 1665591 00:08:40.675 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@650 -- # local es=0 00:08:40.675 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 1665591 00:08:40.675 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:08:40.675 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:40.675 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:08:40.675 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:40.675 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # waitforlisten 1665591 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@831 -- # '[' -z 1665591 ']' 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:40.676 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:08:40.676 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 846: kill: (1665591) - No such process 00:08:40.676 ERROR: process (pid: 1665591) is no longer running 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@864 -- # return 1 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@653 -- # es=1 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:08:40.676 00:08:40.676 real 0m1.275s 00:08:40.676 user 0m1.382s 00:08:40.676 sys 0m0.403s 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:40.676 08:44:30 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:08:40.676 ************************************ 00:08:40.676 END TEST default_locks 00:08:40.676 ************************************ 00:08:40.676 08:44:30 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:08:40.676 08:44:30 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:40.676 08:44:30 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:40.676 08:44:30 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:40.676 ************************************ 00:08:40.676 START TEST default_locks_via_rpc 00:08:40.676 ************************************ 00:08:40.676 08:44:30 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1125 -- # default_locks_via_rpc 00:08:40.676 08:44:30 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=1665825 00:08:40.676 08:44:30 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 1665825 00:08:40.676 08:44:30 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:08:40.676 08:44:30 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 1665825 ']' 00:08:40.676 08:44:30 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:40.676 08:44:30 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:40.676 08:44:30 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:40.676 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:40.676 08:44:30 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:40.676 08:44:30 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:40.676 [2024-11-06 08:44:30.759246] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:40.676 [2024-11-06 08:44:30.759296] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1665825 ] 00:08:40.936 [2024-11-06 08:44:30.829941] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:40.936 [2024-11-06 08:44:30.867441] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 1665825 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 1665825 00:08:41.196 08:44:31 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:41.766 08:44:31 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 1665825 00:08:41.766 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@950 -- # '[' -z 1665825 ']' 00:08:41.766 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # kill -0 1665825 00:08:41.766 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # uname 00:08:41.766 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:41.766 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1665825 00:08:41.767 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:41.767 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:41.767 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1665825' 00:08:41.767 killing process with pid 1665825 00:08:41.767 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@969 -- # kill 1665825 00:08:41.767 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@974 -- # wait 1665825 00:08:41.767 00:08:41.767 real 0m1.158s 00:08:41.767 user 0m1.213s 00:08:41.767 sys 0m0.542s 00:08:41.767 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:41.767 08:44:31 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:41.767 ************************************ 00:08:41.767 END TEST default_locks_via_rpc 00:08:41.767 ************************************ 00:08:42.027 08:44:31 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:08:42.027 08:44:31 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:42.027 08:44:31 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:42.027 08:44:31 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:42.027 ************************************ 00:08:42.027 START TEST non_locking_app_on_locked_coremask 00:08:42.027 ************************************ 00:08:42.027 08:44:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # non_locking_app_on_locked_coremask 00:08:42.027 08:44:31 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:08:42.027 08:44:31 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=1666000 00:08:42.027 08:44:31 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 1666000 /var/tmp/spdk.sock 00:08:42.027 08:44:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 1666000 ']' 00:08:42.027 08:44:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:42.027 08:44:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:42.027 08:44:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:42.027 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:42.027 08:44:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:42.027 08:44:31 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:42.027 [2024-11-06 08:44:31.960573] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:42.027 [2024-11-06 08:44:31.960612] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1666000 ] 00:08:42.027 [2024-11-06 08:44:32.023889] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:42.027 [2024-11-06 08:44:32.060173] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=1666134 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 1666134 /var/tmp/spdk2.sock 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 1666134 ']' 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:42.289 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:42.289 08:44:32 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:42.289 [2024-11-06 08:44:32.311212] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:42.289 [2024-11-06 08:44:32.311269] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1666134 ] 00:08:42.549 [2024-11-06 08:44:32.420942] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:08:42.549 [2024-11-06 08:44:32.420971] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:42.550 [2024-11-06 08:44:32.493358] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:43.120 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:43.120 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:08:43.120 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 1666000 00:08:43.120 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 1666000 00:08:43.120 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:43.691 lslocks: write error 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 1666000 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 1666000 ']' 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 1666000 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1666000 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1666000' 00:08:43.691 killing process with pid 1666000 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 1666000 00:08:43.691 08:44:33 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 1666000 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 1666134 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 1666134 ']' 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 1666134 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1666134 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1666134' 00:08:44.263 killing process with pid 1666134 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 1666134 00:08:44.263 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 1666134 00:08:44.524 00:08:44.524 real 0m2.541s 00:08:44.524 user 0m2.769s 00:08:44.524 sys 0m0.897s 00:08:44.524 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:44.524 08:44:34 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:44.524 ************************************ 00:08:44.524 END TEST non_locking_app_on_locked_coremask 00:08:44.524 ************************************ 00:08:44.525 08:44:34 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:08:44.525 08:44:34 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:44.525 08:44:34 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:44.525 08:44:34 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:44.525 ************************************ 00:08:44.525 START TEST locking_app_on_unlocked_coremask 00:08:44.525 ************************************ 00:08:44.525 08:44:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_unlocked_coremask 00:08:44.525 08:44:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=1666700 00:08:44.525 08:44:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 1666700 /var/tmp/spdk.sock 00:08:44.525 08:44:34 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:08:44.525 08:44:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 1666700 ']' 00:08:44.525 08:44:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:44.525 08:44:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:44.525 08:44:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:44.525 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:44.525 08:44:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:44.525 08:44:34 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:44.525 [2024-11-06 08:44:34.598351] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:44.525 [2024-11-06 08:44:34.598405] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1666700 ] 00:08:44.785 [2024-11-06 08:44:34.668836] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:08:44.785 [2024-11-06 08:44:34.668865] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:44.785 [2024-11-06 08:44:34.706328] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=1666729 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 1666729 /var/tmp/spdk2.sock 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@831 -- # '[' -z 1666729 ']' 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:45.355 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:45.355 08:44:35 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:45.355 [2024-11-06 08:44:35.432143] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:45.355 [2024-11-06 08:44:35.432201] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1666729 ] 00:08:45.616 [2024-11-06 08:44:35.544655] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:45.616 [2024-11-06 08:44:35.617024] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:46.187 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:46.187 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@864 -- # return 0 00:08:46.187 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 1666729 00:08:46.187 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:46.187 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 1666729 00:08:46.757 lslocks: write error 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 1666700 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 1666700 ']' 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 1666700 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1666700 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1666700' 00:08:46.757 killing process with pid 1666700 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 1666700 00:08:46.757 08:44:36 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 1666700 00:08:47.328 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 1666729 00:08:47.328 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@950 -- # '[' -z 1666729 ']' 00:08:47.328 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # kill -0 1666729 00:08:47.328 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # uname 00:08:47.329 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:47.329 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1666729 00:08:47.329 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:47.329 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:47.329 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1666729' 00:08:47.329 killing process with pid 1666729 00:08:47.329 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@969 -- # kill 1666729 00:08:47.329 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@974 -- # wait 1666729 00:08:47.329 00:08:47.329 real 0m2.897s 00:08:47.329 user 0m3.192s 00:08:47.329 sys 0m0.893s 00:08:47.329 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:47.329 08:44:37 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:47.329 ************************************ 00:08:47.329 END TEST locking_app_on_unlocked_coremask 00:08:47.329 ************************************ 00:08:47.589 08:44:37 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:08:47.590 08:44:37 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:47.590 08:44:37 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:47.590 08:44:37 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:47.590 ************************************ 00:08:47.590 START TEST locking_app_on_locked_coremask 00:08:47.590 ************************************ 00:08:47.590 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1125 -- # locking_app_on_locked_coremask 00:08:47.590 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:08:47.590 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=1667322 00:08:47.590 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 1667322 /var/tmp/spdk.sock 00:08:47.590 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 1667322 ']' 00:08:47.590 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:47.590 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:47.590 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:47.590 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:47.590 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:47.590 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:47.590 [2024-11-06 08:44:37.552546] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:47.590 [2024-11-06 08:44:37.552594] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1667322 ] 00:08:47.590 [2024-11-06 08:44:37.621391] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:47.590 [2024-11-06 08:44:37.656867] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 0 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=1667411 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 1667411 /var/tmp/spdk2.sock 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@650 -- # local es=0 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 1667411 /var/tmp/spdk2.sock 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # waitforlisten 1667411 /var/tmp/spdk2.sock 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@831 -- # '[' -z 1667411 ']' 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:47.851 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:47.851 08:44:37 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:47.851 [2024-11-06 08:44:37.901708] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:47.851 [2024-11-06 08:44:37.901769] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1667411 ] 00:08:48.112 [2024-11-06 08:44:38.012236] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 1667322 has claimed it. 00:08:48.112 [2024-11-06 08:44:38.012278] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:08:48.683 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 846: kill: (1667411) - No such process 00:08:48.683 ERROR: process (pid: 1667411) is no longer running 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@864 -- # return 1 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@653 -- # es=1 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 1667322 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 1667322 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:08:48.683 lslocks: write error 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 1667322 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@950 -- # '[' -z 1667322 ']' 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # kill -0 1667322 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # uname 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:48.683 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1667322 00:08:48.944 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:48.944 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:48.944 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1667322' 00:08:48.944 killing process with pid 1667322 00:08:48.944 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@969 -- # kill 1667322 00:08:48.944 08:44:38 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@974 -- # wait 1667322 00:08:48.944 00:08:48.944 real 0m1.534s 00:08:48.944 user 0m1.702s 00:08:48.944 sys 0m0.505s 00:08:48.944 08:44:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:48.944 08:44:39 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:48.944 ************************************ 00:08:48.944 END TEST locking_app_on_locked_coremask 00:08:48.944 ************************************ 00:08:49.205 08:44:39 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:08:49.205 08:44:39 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:49.205 08:44:39 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:49.205 08:44:39 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:49.205 ************************************ 00:08:49.205 START TEST locking_overlapped_coremask 00:08:49.205 ************************************ 00:08:49.205 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask 00:08:49.205 08:44:39 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=1667647 00:08:49.205 08:44:39 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 1667647 /var/tmp/spdk.sock 00:08:49.205 08:44:39 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 00:08:49.205 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 1667647 ']' 00:08:49.205 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:49.205 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:49.205 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:49.205 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:49.205 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:49.205 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:49.205 [2024-11-06 08:44:39.180559] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:49.205 [2024-11-06 08:44:39.180610] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1667647 ] 00:08:49.205 [2024-11-06 08:44:39.253089] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:49.205 [2024-11-06 08:44:39.291364] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:49.205 [2024-11-06 08:44:39.291479] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:08:49.205 [2024-11-06 08:44:39.291484] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 0 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=1667790 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 1667790 /var/tmp/spdk2.sock 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@650 -- # local es=0 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@652 -- # valid_exec_arg waitforlisten 1667790 /var/tmp/spdk2.sock 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@638 -- # local arg=waitforlisten 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # type -t waitforlisten 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # waitforlisten 1667790 /var/tmp/spdk2.sock 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@831 -- # '[' -z 1667790 ']' 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:50.147 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:50.147 08:44:39 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:50.147 [2024-11-06 08:44:40.035592] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:50.147 [2024-11-06 08:44:40.035651] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1667790 ] 00:08:50.147 [2024-11-06 08:44:40.127916] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 1667647 has claimed it. 00:08:50.147 [2024-11-06 08:44:40.127956] app.c: 912:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:08:50.720 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 846: kill: (1667790) - No such process 00:08:50.720 ERROR: process (pid: 1667790) is no longer running 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@864 -- # return 1 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@653 -- # es=1 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 1667647 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@950 -- # '[' -z 1667647 ']' 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # kill -0 1667647 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # uname 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1667647 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1667647' 00:08:50.720 killing process with pid 1667647 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@969 -- # kill 1667647 00:08:50.720 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@974 -- # wait 1667647 00:08:50.981 00:08:50.981 real 0m1.804s 00:08:50.981 user 0m5.221s 00:08:50.981 sys 0m0.389s 00:08:50.981 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:50.981 08:44:40 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:08:50.981 ************************************ 00:08:50.981 END TEST locking_overlapped_coremask 00:08:50.981 ************************************ 00:08:50.981 08:44:40 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:08:50.981 08:44:40 event.cpu_locks -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:50.981 08:44:40 event.cpu_locks -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:50.981 08:44:40 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:50.981 ************************************ 00:08:50.981 START TEST locking_overlapped_coremask_via_rpc 00:08:50.981 ************************************ 00:08:50.981 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1125 -- # locking_overlapped_coremask_via_rpc 00:08:50.982 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=1668134 00:08:50.982 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 1668134 /var/tmp/spdk.sock 00:08:50.982 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:08:50.982 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 1668134 ']' 00:08:50.982 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:50.982 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:50.982 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:50.982 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:50.982 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:50.982 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:50.982 [2024-11-06 08:44:41.064413] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:50.982 [2024-11-06 08:44:41.064469] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1668134 ] 00:08:51.242 [2024-11-06 08:44:41.136828] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:08:51.242 [2024-11-06 08:44:41.136857] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:51.242 [2024-11-06 08:44:41.175099] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:08:51.242 [2024-11-06 08:44:41.175215] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:08:51.242 [2024-11-06 08:44:41.175219] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:51.813 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:51.813 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:08:51.813 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=1668171 00:08:51.814 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:08:51.814 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 1668171 /var/tmp/spdk2.sock 00:08:51.814 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 1668171 ']' 00:08:51.814 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:51.814 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:51.814 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:51.814 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:51.814 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:51.814 08:44:41 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:51.814 [2024-11-06 08:44:41.910799] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:51.814 [2024-11-06 08:44:41.910853] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1668171 ] 00:08:52.075 [2024-11-06 08:44:41.999911] app.c: 916:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:08:52.075 [2024-11-06 08:44:41.999937] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:52.075 [2024-11-06 08:44:42.063449] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:08:52.075 [2024-11-06 08:44:42.063571] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:08:52.075 [2024-11-06 08:44:42.063573] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@650 -- # local es=0 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:52.648 [2024-11-06 08:44:42.730809] app.c: 781:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 1668134 has claimed it. 00:08:52.648 request: 00:08:52.648 { 00:08:52.648 "method": "framework_enable_cpumask_locks", 00:08:52.648 "req_id": 1 00:08:52.648 } 00:08:52.648 Got JSON-RPC error response 00:08:52.648 response: 00:08:52.648 { 00:08:52.648 "code": -32603, 00:08:52.648 "message": "Failed to claim CPU core: 2" 00:08:52.648 } 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@653 -- # es=1 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 1668134 /var/tmp/spdk.sock 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 1668134 ']' 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:52.648 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:52.648 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:52.649 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:52.649 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:52.910 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:52.910 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:08:52.910 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 1668171 /var/tmp/spdk2.sock 00:08:52.910 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@831 -- # '[' -z 1668171 ']' 00:08:52.910 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk2.sock 00:08:52.910 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:52.910 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:08:52.910 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:08:52.910 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:52.910 08:44:42 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:53.175 08:44:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:53.175 08:44:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@864 -- # return 0 00:08:53.175 08:44:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:08:53.175 08:44:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:08:53.175 08:44:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:08:53.175 08:44:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:08:53.175 00:08:53.175 real 0m2.102s 00:08:53.175 user 0m0.863s 00:08:53.175 sys 0m0.152s 00:08:53.175 08:44:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:53.175 08:44:43 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:53.175 ************************************ 00:08:53.175 END TEST locking_overlapped_coremask_via_rpc 00:08:53.175 ************************************ 00:08:53.175 08:44:43 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:08:53.175 08:44:43 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 1668134 ]] 00:08:53.175 08:44:43 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 1668134 00:08:53.175 08:44:43 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 1668134 ']' 00:08:53.175 08:44:43 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 1668134 00:08:53.175 08:44:43 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:08:53.175 08:44:43 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:53.175 08:44:43 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1668134 00:08:53.175 08:44:43 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:53.175 08:44:43 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:53.175 08:44:43 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1668134' 00:08:53.175 killing process with pid 1668134 00:08:53.175 08:44:43 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 1668134 00:08:53.175 08:44:43 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 1668134 00:08:53.437 08:44:43 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 1668171 ]] 00:08:53.437 08:44:43 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 1668171 00:08:53.437 08:44:43 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 1668171 ']' 00:08:53.437 08:44:43 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 1668171 00:08:53.437 08:44:43 event.cpu_locks -- common/autotest_common.sh@955 -- # uname 00:08:53.437 08:44:43 event.cpu_locks -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:53.437 08:44:43 event.cpu_locks -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1668171 00:08:53.437 08:44:43 event.cpu_locks -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:08:53.437 08:44:43 event.cpu_locks -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:08:53.437 08:44:43 event.cpu_locks -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1668171' 00:08:53.437 killing process with pid 1668171 00:08:53.437 08:44:43 event.cpu_locks -- common/autotest_common.sh@969 -- # kill 1668171 00:08:53.437 08:44:43 event.cpu_locks -- common/autotest_common.sh@974 -- # wait 1668171 00:08:53.698 08:44:43 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:08:53.698 08:44:43 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:08:53.698 08:44:43 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 1668134 ]] 00:08:53.698 08:44:43 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 1668134 00:08:53.698 08:44:43 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 1668134 ']' 00:08:53.698 08:44:43 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 1668134 00:08:53.698 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (1668134) - No such process 00:08:53.698 08:44:43 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 1668134 is not found' 00:08:53.698 Process with pid 1668134 is not found 00:08:53.698 08:44:43 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 1668171 ]] 00:08:53.698 08:44:43 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 1668171 00:08:53.698 08:44:43 event.cpu_locks -- common/autotest_common.sh@950 -- # '[' -z 1668171 ']' 00:08:53.698 08:44:43 event.cpu_locks -- common/autotest_common.sh@954 -- # kill -0 1668171 00:08:53.698 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (1668171) - No such process 00:08:53.698 08:44:43 event.cpu_locks -- common/autotest_common.sh@977 -- # echo 'Process with pid 1668171 is not found' 00:08:53.698 Process with pid 1668171 is not found 00:08:53.698 08:44:43 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:08:53.698 00:08:53.698 real 0m14.554s 00:08:53.698 user 0m26.603s 00:08:53.698 sys 0m4.696s 00:08:53.698 08:44:43 event.cpu_locks -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:53.698 08:44:43 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:08:53.698 ************************************ 00:08:53.698 END TEST cpu_locks 00:08:53.698 ************************************ 00:08:53.698 00:08:53.698 real 0m39.387s 00:08:53.698 user 1m17.232s 00:08:53.698 sys 0m7.871s 00:08:53.698 08:44:43 event -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:53.698 08:44:43 event -- common/autotest_common.sh@10 -- # set +x 00:08:53.698 ************************************ 00:08:53.698 END TEST event 00:08:53.698 ************************************ 00:08:53.698 08:44:43 -- spdk/autotest.sh@169 -- # run_test thread /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:08:53.698 08:44:43 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:53.698 08:44:43 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:53.698 08:44:43 -- common/autotest_common.sh@10 -- # set +x 00:08:53.698 ************************************ 00:08:53.698 START TEST thread 00:08:53.698 ************************************ 00:08:53.698 08:44:43 thread -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:08:53.959 * Looking for test storage... 00:08:53.959 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread 00:08:53.959 08:44:43 thread -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:53.959 08:44:43 thread -- common/autotest_common.sh@1689 -- # lcov --version 00:08:53.959 08:44:43 thread -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:53.959 08:44:43 thread -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:53.959 08:44:43 thread -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:53.959 08:44:43 thread -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:53.959 08:44:43 thread -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:53.959 08:44:43 thread -- scripts/common.sh@336 -- # IFS=.-: 00:08:53.959 08:44:43 thread -- scripts/common.sh@336 -- # read -ra ver1 00:08:53.959 08:44:43 thread -- scripts/common.sh@337 -- # IFS=.-: 00:08:53.959 08:44:43 thread -- scripts/common.sh@337 -- # read -ra ver2 00:08:53.959 08:44:43 thread -- scripts/common.sh@338 -- # local 'op=<' 00:08:53.959 08:44:43 thread -- scripts/common.sh@340 -- # ver1_l=2 00:08:53.959 08:44:43 thread -- scripts/common.sh@341 -- # ver2_l=1 00:08:53.959 08:44:43 thread -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:53.960 08:44:43 thread -- scripts/common.sh@344 -- # case "$op" in 00:08:53.960 08:44:43 thread -- scripts/common.sh@345 -- # : 1 00:08:53.960 08:44:43 thread -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:53.960 08:44:43 thread -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:53.960 08:44:43 thread -- scripts/common.sh@365 -- # decimal 1 00:08:53.960 08:44:43 thread -- scripts/common.sh@353 -- # local d=1 00:08:53.960 08:44:43 thread -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:53.960 08:44:43 thread -- scripts/common.sh@355 -- # echo 1 00:08:53.960 08:44:43 thread -- scripts/common.sh@365 -- # ver1[v]=1 00:08:53.960 08:44:43 thread -- scripts/common.sh@366 -- # decimal 2 00:08:53.960 08:44:43 thread -- scripts/common.sh@353 -- # local d=2 00:08:53.960 08:44:43 thread -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:53.960 08:44:43 thread -- scripts/common.sh@355 -- # echo 2 00:08:53.960 08:44:44 thread -- scripts/common.sh@366 -- # ver2[v]=2 00:08:53.960 08:44:44 thread -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:53.960 08:44:44 thread -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:53.960 08:44:44 thread -- scripts/common.sh@368 -- # return 0 00:08:53.960 08:44:44 thread -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:53.960 08:44:44 thread -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:53.960 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:53.960 --rc genhtml_branch_coverage=1 00:08:53.960 --rc genhtml_function_coverage=1 00:08:53.960 --rc genhtml_legend=1 00:08:53.960 --rc geninfo_all_blocks=1 00:08:53.960 --rc geninfo_unexecuted_blocks=1 00:08:53.960 00:08:53.960 ' 00:08:53.960 08:44:44 thread -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:53.960 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:53.960 --rc genhtml_branch_coverage=1 00:08:53.960 --rc genhtml_function_coverage=1 00:08:53.960 --rc genhtml_legend=1 00:08:53.960 --rc geninfo_all_blocks=1 00:08:53.960 --rc geninfo_unexecuted_blocks=1 00:08:53.960 00:08:53.960 ' 00:08:53.960 08:44:44 thread -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:53.960 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:53.960 --rc genhtml_branch_coverage=1 00:08:53.960 --rc genhtml_function_coverage=1 00:08:53.960 --rc genhtml_legend=1 00:08:53.960 --rc geninfo_all_blocks=1 00:08:53.960 --rc geninfo_unexecuted_blocks=1 00:08:53.960 00:08:53.960 ' 00:08:53.960 08:44:44 thread -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:53.960 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:53.960 --rc genhtml_branch_coverage=1 00:08:53.960 --rc genhtml_function_coverage=1 00:08:53.960 --rc genhtml_legend=1 00:08:53.960 --rc geninfo_all_blocks=1 00:08:53.960 --rc geninfo_unexecuted_blocks=1 00:08:53.960 00:08:53.960 ' 00:08:53.960 08:44:44 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:08:53.960 08:44:44 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:08:53.960 08:44:44 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:53.960 08:44:44 thread -- common/autotest_common.sh@10 -- # set +x 00:08:53.960 ************************************ 00:08:53.960 START TEST thread_poller_perf 00:08:53.960 ************************************ 00:08:53.960 08:44:44 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:08:53.960 [2024-11-06 08:44:44.064940] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:53.960 [2024-11-06 08:44:44.065045] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1668748 ] 00:08:54.220 [2024-11-06 08:44:44.142931] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:54.220 [2024-11-06 08:44:44.184807] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:54.220 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:08:55.161 [2024-11-06T07:44:45.274Z] ====================================== 00:08:55.161 [2024-11-06T07:44:45.274Z] busy:2408776618 (cyc) 00:08:55.161 [2024-11-06T07:44:45.274Z] total_run_count: 288000 00:08:55.161 [2024-11-06T07:44:45.274Z] tsc_hz: 2400000000 (cyc) 00:08:55.161 [2024-11-06T07:44:45.274Z] ====================================== 00:08:55.161 [2024-11-06T07:44:45.274Z] poller_cost: 8363 (cyc), 3484 (nsec) 00:08:55.161 00:08:55.161 real 0m1.182s 00:08:55.161 user 0m1.111s 00:08:55.161 sys 0m0.067s 00:08:55.161 08:44:45 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:55.161 08:44:45 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:08:55.161 ************************************ 00:08:55.162 END TEST thread_poller_perf 00:08:55.162 ************************************ 00:08:55.162 08:44:45 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:08:55.162 08:44:45 thread -- common/autotest_common.sh@1101 -- # '[' 8 -le 1 ']' 00:08:55.162 08:44:45 thread -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:55.162 08:44:45 thread -- common/autotest_common.sh@10 -- # set +x 00:08:55.422 ************************************ 00:08:55.422 START TEST thread_poller_perf 00:08:55.422 ************************************ 00:08:55.422 08:44:45 thread.thread_poller_perf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:08:55.422 [2024-11-06 08:44:45.323570] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:55.422 [2024-11-06 08:44:45.323674] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1668968 ] 00:08:55.422 [2024-11-06 08:44:45.399381] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:55.422 [2024-11-06 08:44:45.437279] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:55.422 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:08:56.366 [2024-11-06T07:44:46.479Z] ====================================== 00:08:56.366 [2024-11-06T07:44:46.479Z] busy:2401838068 (cyc) 00:08:56.366 [2024-11-06T07:44:46.479Z] total_run_count: 3813000 00:08:56.366 [2024-11-06T07:44:46.479Z] tsc_hz: 2400000000 (cyc) 00:08:56.366 [2024-11-06T07:44:46.479Z] ====================================== 00:08:56.366 [2024-11-06T07:44:46.479Z] poller_cost: 629 (cyc), 262 (nsec) 00:08:56.366 00:08:56.366 real 0m1.167s 00:08:56.366 user 0m1.095s 00:08:56.366 sys 0m0.068s 00:08:56.366 08:44:46 thread.thread_poller_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:56.366 08:44:46 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:08:56.366 ************************************ 00:08:56.366 END TEST thread_poller_perf 00:08:56.366 ************************************ 00:08:56.628 08:44:46 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:08:56.628 00:08:56.628 real 0m2.707s 00:08:56.628 user 0m2.375s 00:08:56.628 sys 0m0.345s 00:08:56.628 08:44:46 thread -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:56.628 08:44:46 thread -- common/autotest_common.sh@10 -- # set +x 00:08:56.628 ************************************ 00:08:56.628 END TEST thread 00:08:56.628 ************************************ 00:08:56.628 08:44:46 -- spdk/autotest.sh@171 -- # [[ 0 -eq 1 ]] 00:08:56.628 08:44:46 -- spdk/autotest.sh@176 -- # run_test app_cmdline /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:08:56.628 08:44:46 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:56.628 08:44:46 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:56.628 08:44:46 -- common/autotest_common.sh@10 -- # set +x 00:08:56.628 ************************************ 00:08:56.628 START TEST app_cmdline 00:08:56.628 ************************************ 00:08:56.628 08:44:46 app_cmdline -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:08:56.628 * Looking for test storage... 00:08:56.628 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:08:56.628 08:44:46 app_cmdline -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:56.628 08:44:46 app_cmdline -- common/autotest_common.sh@1689 -- # lcov --version 00:08:56.628 08:44:46 app_cmdline -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@336 -- # IFS=.-: 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@336 -- # read -ra ver1 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@337 -- # IFS=.-: 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@337 -- # read -ra ver2 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@338 -- # local 'op=<' 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@340 -- # ver1_l=2 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@341 -- # ver2_l=1 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@344 -- # case "$op" in 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@345 -- # : 1 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@365 -- # decimal 1 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@353 -- # local d=1 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@355 -- # echo 1 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@365 -- # ver1[v]=1 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@366 -- # decimal 2 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@353 -- # local d=2 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@355 -- # echo 2 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@366 -- # ver2[v]=2 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:56.889 08:44:46 app_cmdline -- scripts/common.sh@368 -- # return 0 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:56.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:56.889 --rc genhtml_branch_coverage=1 00:08:56.889 --rc genhtml_function_coverage=1 00:08:56.889 --rc genhtml_legend=1 00:08:56.889 --rc geninfo_all_blocks=1 00:08:56.889 --rc geninfo_unexecuted_blocks=1 00:08:56.889 00:08:56.889 ' 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:56.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:56.889 --rc genhtml_branch_coverage=1 00:08:56.889 --rc genhtml_function_coverage=1 00:08:56.889 --rc genhtml_legend=1 00:08:56.889 --rc geninfo_all_blocks=1 00:08:56.889 --rc geninfo_unexecuted_blocks=1 00:08:56.889 00:08:56.889 ' 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:56.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:56.889 --rc genhtml_branch_coverage=1 00:08:56.889 --rc genhtml_function_coverage=1 00:08:56.889 --rc genhtml_legend=1 00:08:56.889 --rc geninfo_all_blocks=1 00:08:56.889 --rc geninfo_unexecuted_blocks=1 00:08:56.889 00:08:56.889 ' 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:56.889 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:56.889 --rc genhtml_branch_coverage=1 00:08:56.889 --rc genhtml_function_coverage=1 00:08:56.889 --rc genhtml_legend=1 00:08:56.889 --rc geninfo_all_blocks=1 00:08:56.889 --rc geninfo_unexecuted_blocks=1 00:08:56.889 00:08:56.889 ' 00:08:56.889 08:44:46 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:08:56.889 08:44:46 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=1669375 00:08:56.889 08:44:46 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 1669375 00:08:56.889 08:44:46 app_cmdline -- app/cmdline.sh@16 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@831 -- # '[' -z 1669375 ']' 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@836 -- # local max_retries=100 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:56.889 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@840 -- # xtrace_disable 00:08:56.889 08:44:46 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:56.889 [2024-11-06 08:44:46.842218] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:08:56.889 [2024-11-06 08:44:46.842275] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1669375 ] 00:08:56.889 [2024-11-06 08:44:46.913555] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:08:56.889 [2024-11-06 08:44:46.949431] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@864 -- # return 0 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py spdk_get_version 00:08:57.831 { 00:08:57.831 "version": "SPDK v25.01-pre git sha1 15b283ee8", 00:08:57.831 "fields": { 00:08:57.831 "major": 25, 00:08:57.831 "minor": 1, 00:08:57.831 "patch": 0, 00:08:57.831 "suffix": "-pre", 00:08:57.831 "commit": "15b283ee8" 00:08:57.831 } 00:08:57.831 } 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@26 -- # sort 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@561 -- # xtrace_disable 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:08:57.831 08:44:47 app_cmdline -- app/cmdline.sh@30 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@650 -- # local es=0 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:08:57.831 08:44:47 app_cmdline -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:08:58.091 request: 00:08:58.091 { 00:08:58.091 "method": "env_dpdk_get_mem_stats", 00:08:58.091 "req_id": 1 00:08:58.091 } 00:08:58.091 Got JSON-RPC error response 00:08:58.091 response: 00:08:58.091 { 00:08:58.091 "code": -32601, 00:08:58.091 "message": "Method not found" 00:08:58.091 } 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@653 -- # es=1 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:08:58.091 08:44:48 app_cmdline -- app/cmdline.sh@1 -- # killprocess 1669375 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@950 -- # '[' -z 1669375 ']' 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@954 -- # kill -0 1669375 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@955 -- # uname 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1669375 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1669375' 00:08:58.091 killing process with pid 1669375 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@969 -- # kill 1669375 00:08:58.091 08:44:48 app_cmdline -- common/autotest_common.sh@974 -- # wait 1669375 00:08:58.351 00:08:58.351 real 0m1.744s 00:08:58.351 user 0m2.104s 00:08:58.351 sys 0m0.444s 00:08:58.351 08:44:48 app_cmdline -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:58.351 08:44:48 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:08:58.351 ************************************ 00:08:58.351 END TEST app_cmdline 00:08:58.351 ************************************ 00:08:58.351 08:44:48 -- spdk/autotest.sh@177 -- # run_test version /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:08:58.351 08:44:48 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:08:58.351 08:44:48 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:58.351 08:44:48 -- common/autotest_common.sh@10 -- # set +x 00:08:58.351 ************************************ 00:08:58.351 START TEST version 00:08:58.351 ************************************ 00:08:58.351 08:44:48 version -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:08:58.612 * Looking for test storage... 00:08:58.612 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:08:58.612 08:44:48 version -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:58.612 08:44:48 version -- common/autotest_common.sh@1689 -- # lcov --version 00:08:58.612 08:44:48 version -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:58.612 08:44:48 version -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:58.612 08:44:48 version -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:58.612 08:44:48 version -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:58.612 08:44:48 version -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:58.612 08:44:48 version -- scripts/common.sh@336 -- # IFS=.-: 00:08:58.612 08:44:48 version -- scripts/common.sh@336 -- # read -ra ver1 00:08:58.612 08:44:48 version -- scripts/common.sh@337 -- # IFS=.-: 00:08:58.612 08:44:48 version -- scripts/common.sh@337 -- # read -ra ver2 00:08:58.612 08:44:48 version -- scripts/common.sh@338 -- # local 'op=<' 00:08:58.612 08:44:48 version -- scripts/common.sh@340 -- # ver1_l=2 00:08:58.612 08:44:48 version -- scripts/common.sh@341 -- # ver2_l=1 00:08:58.612 08:44:48 version -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:58.612 08:44:48 version -- scripts/common.sh@344 -- # case "$op" in 00:08:58.612 08:44:48 version -- scripts/common.sh@345 -- # : 1 00:08:58.612 08:44:48 version -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:58.612 08:44:48 version -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:58.612 08:44:48 version -- scripts/common.sh@365 -- # decimal 1 00:08:58.612 08:44:48 version -- scripts/common.sh@353 -- # local d=1 00:08:58.612 08:44:48 version -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:58.612 08:44:48 version -- scripts/common.sh@355 -- # echo 1 00:08:58.612 08:44:48 version -- scripts/common.sh@365 -- # ver1[v]=1 00:08:58.612 08:44:48 version -- scripts/common.sh@366 -- # decimal 2 00:08:58.612 08:44:48 version -- scripts/common.sh@353 -- # local d=2 00:08:58.612 08:44:48 version -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:58.612 08:44:48 version -- scripts/common.sh@355 -- # echo 2 00:08:58.612 08:44:48 version -- scripts/common.sh@366 -- # ver2[v]=2 00:08:58.612 08:44:48 version -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:58.612 08:44:48 version -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:58.612 08:44:48 version -- scripts/common.sh@368 -- # return 0 00:08:58.612 08:44:48 version -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:58.612 08:44:48 version -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:58.612 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:58.612 --rc genhtml_branch_coverage=1 00:08:58.612 --rc genhtml_function_coverage=1 00:08:58.612 --rc genhtml_legend=1 00:08:58.612 --rc geninfo_all_blocks=1 00:08:58.612 --rc geninfo_unexecuted_blocks=1 00:08:58.612 00:08:58.612 ' 00:08:58.612 08:44:48 version -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:58.612 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:58.612 --rc genhtml_branch_coverage=1 00:08:58.612 --rc genhtml_function_coverage=1 00:08:58.612 --rc genhtml_legend=1 00:08:58.612 --rc geninfo_all_blocks=1 00:08:58.613 --rc geninfo_unexecuted_blocks=1 00:08:58.613 00:08:58.613 ' 00:08:58.613 08:44:48 version -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:58.613 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:58.613 --rc genhtml_branch_coverage=1 00:08:58.613 --rc genhtml_function_coverage=1 00:08:58.613 --rc genhtml_legend=1 00:08:58.613 --rc geninfo_all_blocks=1 00:08:58.613 --rc geninfo_unexecuted_blocks=1 00:08:58.613 00:08:58.613 ' 00:08:58.613 08:44:48 version -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:58.613 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:58.613 --rc genhtml_branch_coverage=1 00:08:58.613 --rc genhtml_function_coverage=1 00:08:58.613 --rc genhtml_legend=1 00:08:58.613 --rc geninfo_all_blocks=1 00:08:58.613 --rc geninfo_unexecuted_blocks=1 00:08:58.613 00:08:58.613 ' 00:08:58.613 08:44:48 version -- app/version.sh@17 -- # get_header_version major 00:08:58.613 08:44:48 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:08:58.613 08:44:48 version -- app/version.sh@14 -- # cut -f2 00:08:58.613 08:44:48 version -- app/version.sh@14 -- # tr -d '"' 00:08:58.613 08:44:48 version -- app/version.sh@17 -- # major=25 00:08:58.613 08:44:48 version -- app/version.sh@18 -- # get_header_version minor 00:08:58.613 08:44:48 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:08:58.613 08:44:48 version -- app/version.sh@14 -- # cut -f2 00:08:58.613 08:44:48 version -- app/version.sh@14 -- # tr -d '"' 00:08:58.613 08:44:48 version -- app/version.sh@18 -- # minor=1 00:08:58.613 08:44:48 version -- app/version.sh@19 -- # get_header_version patch 00:08:58.613 08:44:48 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:08:58.613 08:44:48 version -- app/version.sh@14 -- # cut -f2 00:08:58.613 08:44:48 version -- app/version.sh@14 -- # tr -d '"' 00:08:58.613 08:44:48 version -- app/version.sh@19 -- # patch=0 00:08:58.613 08:44:48 version -- app/version.sh@20 -- # get_header_version suffix 00:08:58.613 08:44:48 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:08:58.613 08:44:48 version -- app/version.sh@14 -- # cut -f2 00:08:58.613 08:44:48 version -- app/version.sh@14 -- # tr -d '"' 00:08:58.613 08:44:48 version -- app/version.sh@20 -- # suffix=-pre 00:08:58.613 08:44:48 version -- app/version.sh@22 -- # version=25.1 00:08:58.613 08:44:48 version -- app/version.sh@25 -- # (( patch != 0 )) 00:08:58.613 08:44:48 version -- app/version.sh@28 -- # version=25.1rc0 00:08:58.613 08:44:48 version -- app/version.sh@30 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:08:58.613 08:44:48 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:08:58.613 08:44:48 version -- app/version.sh@30 -- # py_version=25.1rc0 00:08:58.613 08:44:48 version -- app/version.sh@31 -- # [[ 25.1rc0 == \2\5\.\1\r\c\0 ]] 00:08:58.613 00:08:58.613 real 0m0.257s 00:08:58.613 user 0m0.151s 00:08:58.613 sys 0m0.146s 00:08:58.613 08:44:48 version -- common/autotest_common.sh@1126 -- # xtrace_disable 00:08:58.613 08:44:48 version -- common/autotest_common.sh@10 -- # set +x 00:08:58.613 ************************************ 00:08:58.613 END TEST version 00:08:58.613 ************************************ 00:08:58.613 08:44:48 -- spdk/autotest.sh@179 -- # '[' 0 -eq 1 ']' 00:08:58.613 08:44:48 -- spdk/autotest.sh@188 -- # [[ 0 -eq 1 ]] 00:08:58.613 08:44:48 -- spdk/autotest.sh@194 -- # uname -s 00:08:58.613 08:44:48 -- spdk/autotest.sh@194 -- # [[ Linux == Linux ]] 00:08:58.613 08:44:48 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:08:58.613 08:44:48 -- spdk/autotest.sh@195 -- # [[ 0 -eq 1 ]] 00:08:58.613 08:44:48 -- spdk/autotest.sh@207 -- # '[' 0 -eq 1 ']' 00:08:58.613 08:44:48 -- spdk/autotest.sh@252 -- # '[' 0 -eq 1 ']' 00:08:58.613 08:44:48 -- spdk/autotest.sh@256 -- # timing_exit lib 00:08:58.613 08:44:48 -- common/autotest_common.sh@730 -- # xtrace_disable 00:08:58.613 08:44:48 -- common/autotest_common.sh@10 -- # set +x 00:08:58.874 08:44:48 -- spdk/autotest.sh@258 -- # '[' 0 -eq 1 ']' 00:08:58.874 08:44:48 -- spdk/autotest.sh@263 -- # '[' 0 -eq 1 ']' 00:08:58.874 08:44:48 -- spdk/autotest.sh@272 -- # '[' 1 -eq 1 ']' 00:08:58.874 08:44:48 -- spdk/autotest.sh@273 -- # export NET_TYPE 00:08:58.874 08:44:48 -- spdk/autotest.sh@276 -- # '[' tcp = rdma ']' 00:08:58.874 08:44:48 -- spdk/autotest.sh@279 -- # '[' tcp = tcp ']' 00:08:58.874 08:44:48 -- spdk/autotest.sh@280 -- # run_test nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:08:58.874 08:44:48 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:08:58.874 08:44:48 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:58.874 08:44:48 -- common/autotest_common.sh@10 -- # set +x 00:08:58.874 ************************************ 00:08:58.874 START TEST nvmf_tcp 00:08:58.874 ************************************ 00:08:58.874 08:44:48 nvmf_tcp -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:08:58.874 * Looking for test storage... 00:08:58.874 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:08:58.874 08:44:48 nvmf_tcp -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:58.874 08:44:48 nvmf_tcp -- common/autotest_common.sh@1689 -- # lcov --version 00:08:58.874 08:44:48 nvmf_tcp -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:58.874 08:44:48 nvmf_tcp -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@344 -- # case "$op" in 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@345 -- # : 1 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@365 -- # decimal 1 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@353 -- # local d=1 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@355 -- # echo 1 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@366 -- # decimal 2 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@353 -- # local d=2 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@355 -- # echo 2 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:58.874 08:44:48 nvmf_tcp -- scripts/common.sh@368 -- # return 0 00:08:58.874 08:44:48 nvmf_tcp -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:58.874 08:44:48 nvmf_tcp -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:58.874 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:58.874 --rc genhtml_branch_coverage=1 00:08:58.874 --rc genhtml_function_coverage=1 00:08:58.874 --rc genhtml_legend=1 00:08:58.874 --rc geninfo_all_blocks=1 00:08:58.874 --rc geninfo_unexecuted_blocks=1 00:08:58.874 00:08:58.874 ' 00:08:58.874 08:44:48 nvmf_tcp -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:58.874 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:58.874 --rc genhtml_branch_coverage=1 00:08:58.874 --rc genhtml_function_coverage=1 00:08:58.874 --rc genhtml_legend=1 00:08:58.874 --rc geninfo_all_blocks=1 00:08:58.874 --rc geninfo_unexecuted_blocks=1 00:08:58.874 00:08:58.874 ' 00:08:58.874 08:44:48 nvmf_tcp -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:58.874 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:58.874 --rc genhtml_branch_coverage=1 00:08:58.874 --rc genhtml_function_coverage=1 00:08:58.874 --rc genhtml_legend=1 00:08:58.874 --rc geninfo_all_blocks=1 00:08:58.874 --rc geninfo_unexecuted_blocks=1 00:08:58.874 00:08:58.874 ' 00:08:58.874 08:44:48 nvmf_tcp -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:58.874 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:58.874 --rc genhtml_branch_coverage=1 00:08:58.874 --rc genhtml_function_coverage=1 00:08:58.874 --rc genhtml_legend=1 00:08:58.874 --rc geninfo_all_blocks=1 00:08:58.874 --rc geninfo_unexecuted_blocks=1 00:08:58.874 00:08:58.874 ' 00:08:58.874 08:44:48 nvmf_tcp -- nvmf/nvmf.sh@10 -- # uname -s 00:08:59.135 08:44:48 nvmf_tcp -- nvmf/nvmf.sh@10 -- # '[' '!' Linux = Linux ']' 00:08:59.135 08:44:48 nvmf_tcp -- nvmf/nvmf.sh@14 -- # run_test nvmf_target_core /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp 00:08:59.135 08:44:48 nvmf_tcp -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:08:59.135 08:44:48 nvmf_tcp -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:59.135 08:44:48 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:08:59.135 ************************************ 00:08:59.135 START TEST nvmf_target_core 00:08:59.135 ************************************ 00:08:59.135 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp 00:08:59.135 * Looking for test storage... 00:08:59.135 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:08:59.135 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:59.135 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1689 -- # lcov --version 00:08:59.135 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:59.135 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:59.135 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:59.135 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:59.135 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:59.135 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@336 -- # IFS=.-: 00:08:59.135 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@336 -- # read -ra ver1 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@337 -- # IFS=.-: 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@337 -- # read -ra ver2 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@338 -- # local 'op=<' 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@340 -- # ver1_l=2 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@341 -- # ver2_l=1 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@344 -- # case "$op" in 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@345 -- # : 1 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@365 -- # decimal 1 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@353 -- # local d=1 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@355 -- # echo 1 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@365 -- # ver1[v]=1 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@366 -- # decimal 2 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@353 -- # local d=2 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@355 -- # echo 2 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@366 -- # ver2[v]=2 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@368 -- # return 0 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:59.136 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:59.136 --rc genhtml_branch_coverage=1 00:08:59.136 --rc genhtml_function_coverage=1 00:08:59.136 --rc genhtml_legend=1 00:08:59.136 --rc geninfo_all_blocks=1 00:08:59.136 --rc geninfo_unexecuted_blocks=1 00:08:59.136 00:08:59.136 ' 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:59.136 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:59.136 --rc genhtml_branch_coverage=1 00:08:59.136 --rc genhtml_function_coverage=1 00:08:59.136 --rc genhtml_legend=1 00:08:59.136 --rc geninfo_all_blocks=1 00:08:59.136 --rc geninfo_unexecuted_blocks=1 00:08:59.136 00:08:59.136 ' 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:59.136 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:59.136 --rc genhtml_branch_coverage=1 00:08:59.136 --rc genhtml_function_coverage=1 00:08:59.136 --rc genhtml_legend=1 00:08:59.136 --rc geninfo_all_blocks=1 00:08:59.136 --rc geninfo_unexecuted_blocks=1 00:08:59.136 00:08:59.136 ' 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:59.136 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:59.136 --rc genhtml_branch_coverage=1 00:08:59.136 --rc genhtml_function_coverage=1 00:08:59.136 --rc genhtml_legend=1 00:08:59.136 --rc geninfo_all_blocks=1 00:08:59.136 --rc geninfo_unexecuted_blocks=1 00:08:59.136 00:08:59.136 ' 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@10 -- # uname -s 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@10 -- # '[' '!' Linux = Linux ']' 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@7 -- # uname -s 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:59.136 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:08:59.398 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:08:59.398 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:59.398 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:59.398 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:59.398 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:59.398 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:59.398 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@15 -- # shopt -s extglob 00:08:59.398 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:59.398 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:59.398 08:44:49 nvmf_tcp.nvmf_target_core -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- paths/export.sh@5 -- # export PATH 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@51 -- # : 0 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:08:59.399 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/common.sh@55 -- # have_pci_nics=0 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@16 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@18 -- # TEST_ARGS=("$@") 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@20 -- # [[ 0 -eq 0 ]] 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@21 -- # run_test nvmf_abort /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:08:59.399 ************************************ 00:08:59.399 START TEST nvmf_abort 00:08:59.399 ************************************ 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:08:59.399 * Looking for test storage... 00:08:59.399 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1689 -- # lcov --version 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@333 -- # local ver1 ver1_l 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@334 -- # local ver2 ver2_l 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@336 -- # IFS=.-: 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@336 -- # read -ra ver1 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@337 -- # IFS=.-: 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@337 -- # read -ra ver2 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@338 -- # local 'op=<' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@340 -- # ver1_l=2 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@341 -- # ver2_l=1 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@344 -- # case "$op" in 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@345 -- # : 1 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@364 -- # (( v = 0 )) 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@365 -- # decimal 1 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@353 -- # local d=1 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@355 -- # echo 1 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@365 -- # ver1[v]=1 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@366 -- # decimal 2 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@353 -- # local d=2 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@355 -- # echo 2 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@366 -- # ver2[v]=2 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@368 -- # return 0 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:08:59.399 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:59.399 --rc genhtml_branch_coverage=1 00:08:59.399 --rc genhtml_function_coverage=1 00:08:59.399 --rc genhtml_legend=1 00:08:59.399 --rc geninfo_all_blocks=1 00:08:59.399 --rc geninfo_unexecuted_blocks=1 00:08:59.399 00:08:59.399 ' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:08:59.399 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:59.399 --rc genhtml_branch_coverage=1 00:08:59.399 --rc genhtml_function_coverage=1 00:08:59.399 --rc genhtml_legend=1 00:08:59.399 --rc geninfo_all_blocks=1 00:08:59.399 --rc geninfo_unexecuted_blocks=1 00:08:59.399 00:08:59.399 ' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:08:59.399 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:59.399 --rc genhtml_branch_coverage=1 00:08:59.399 --rc genhtml_function_coverage=1 00:08:59.399 --rc genhtml_legend=1 00:08:59.399 --rc geninfo_all_blocks=1 00:08:59.399 --rc geninfo_unexecuted_blocks=1 00:08:59.399 00:08:59.399 ' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:08:59.399 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:08:59.399 --rc genhtml_branch_coverage=1 00:08:59.399 --rc genhtml_function_coverage=1 00:08:59.399 --rc genhtml_legend=1 00:08:59.399 --rc geninfo_all_blocks=1 00:08:59.399 --rc geninfo_unexecuted_blocks=1 00:08:59.399 00:08:59.399 ' 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@7 -- # uname -s 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:59.399 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:59.661 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:59.661 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:08:59.661 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:08:59.661 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:59.661 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@15 -- # shopt -s extglob 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@5 -- # export PATH 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@51 -- # : 0 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:08:59.662 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@55 -- # have_pci_nics=0 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@11 -- # MALLOC_BDEV_SIZE=64 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@12 -- # MALLOC_BLOCK_SIZE=4096 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@14 -- # nvmftestinit 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@474 -- # prepare_net_devs 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@436 -- # local -g is_hw=no 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@438 -- # remove_spdk_ns 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@309 -- # xtrace_disable 00:08:59.662 08:44:49 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@315 -- # pci_devs=() 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@315 -- # local -a pci_devs 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@316 -- # pci_net_devs=() 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@317 -- # pci_drivers=() 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@317 -- # local -A pci_drivers 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@319 -- # net_devs=() 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@319 -- # local -ga net_devs 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@320 -- # e810=() 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@320 -- # local -ga e810 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@321 -- # x722=() 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@321 -- # local -ga x722 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@322 -- # mlx=() 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@322 -- # local -ga mlx 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:09:07.807 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:09:07.807 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@416 -- # [[ up == up ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:09:07.807 Found net devices under 0000:4b:00.0: cvl_0_0 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@416 -- # [[ up == up ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:09:07.807 Found net devices under 0000:4b:00.1: cvl_0_1 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@440 -- # is_hw=yes 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:09:07.807 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:07.807 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.636 ms 00:09:07.807 00:09:07.807 --- 10.0.0.2 ping statistics --- 00:09:07.807 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:07.807 rtt min/avg/max/mdev = 0.636/0.636/0.636/0.000 ms 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:07.807 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:07.807 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.284 ms 00:09:07.807 00:09:07.807 --- 10.0.0.1 ping statistics --- 00:09:07.807 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:07.807 rtt min/avg/max/mdev = 0.284/0.284/0.284/0.000 ms 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@448 -- # return 0 00:09:07.807 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@15 -- # nvmfappstart -m 0xE 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@724 -- # xtrace_disable 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@507 -- # nvmfpid=1673862 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@508 -- # waitforlisten 1673862 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@831 -- # '[' -z 1673862 ']' 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:07.808 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:07.808 08:44:56 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:07.808 [2024-11-06 08:44:56.978859] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:09:07.808 [2024-11-06 08:44:56.978910] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:07.808 [2024-11-06 08:44:57.074430] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:09:07.808 [2024-11-06 08:44:57.120231] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:09:07.808 [2024-11-06 08:44:57.120282] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:09:07.808 [2024-11-06 08:44:57.120291] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:09:07.808 [2024-11-06 08:44:57.120298] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:09:07.808 [2024-11-06 08:44:57.120304] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:09:07.808 [2024-11-06 08:44:57.122011] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:09:07.808 [2024-11-06 08:44:57.122178] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:07.808 [2024-11-06 08:44:57.122179] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@864 -- # return 0 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@730 -- # xtrace_disable 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -a 256 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:07.808 [2024-11-06 08:44:57.831240] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@20 -- # rpc_cmd bdev_malloc_create 64 4096 -b Malloc0 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:07.808 Malloc0 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@21 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:07.808 Delay0 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 Delay0 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:07.808 [2024-11-06 08:44:57.911993] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:07.808 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:08.069 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:08.069 08:44:57 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0x1 -t 1 -l warning -q 128 00:09:08.069 [2024-11-06 08:44:58.043115] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:09:10.612 Initializing NVMe Controllers 00:09:10.612 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:09:10.612 controller IO queue size 128 less than required 00:09:10.612 Consider using lower queue depth or small IO size because IO requests may be queued at the NVMe driver. 00:09:10.612 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 0 00:09:10.612 Initialization complete. Launching workers. 00:09:10.612 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 I/O completed: 127, failed: 29030 00:09:10.612 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) abort submitted 29095, failed to submit 62 00:09:10.612 success 29034, unsuccessful 61, failed 0 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- target/abort.sh@38 -- # nvmftestfini 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@514 -- # nvmfcleanup 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@121 -- # sync 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@124 -- # set +e 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@125 -- # for i in {1..20} 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:09:10.612 rmmod nvme_tcp 00:09:10.612 rmmod nvme_fabrics 00:09:10.612 rmmod nvme_keyring 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@128 -- # set -e 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@129 -- # return 0 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@515 -- # '[' -n 1673862 ']' 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@516 -- # killprocess 1673862 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@950 -- # '[' -z 1673862 ']' 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@954 -- # kill -0 1673862 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@955 -- # uname 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1673862 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1673862' 00:09:10.612 killing process with pid 1673862 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@969 -- # kill 1673862 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@974 -- # wait 1673862 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@297 -- # iptr 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@789 -- # iptables-save 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@789 -- # iptables-restore 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@302 -- # remove_spdk_ns 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:10.612 08:45:00 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:12.523 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_abort -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:09:12.523 00:09:12.523 real 0m13.180s 00:09:12.523 user 0m13.883s 00:09:12.523 sys 0m6.319s 00:09:12.523 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@1126 -- # xtrace_disable 00:09:12.523 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:09:12.523 ************************************ 00:09:12.523 END TEST nvmf_abort 00:09:12.523 ************************************ 00:09:12.523 08:45:02 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@22 -- # run_test nvmf_ns_hotplug_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:09:12.523 08:45:02 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:09:12.523 08:45:02 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:09:12.523 08:45:02 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:09:12.523 ************************************ 00:09:12.523 START TEST nvmf_ns_hotplug_stress 00:09:12.523 ************************************ 00:09:12.523 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:09:12.784 * Looking for test storage... 00:09:12.784 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1689 -- # lcov --version 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # IFS=.-: 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # read -ra ver1 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # IFS=.-: 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # read -ra ver2 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@338 -- # local 'op=<' 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@340 -- # ver1_l=2 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@341 -- # ver2_l=1 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@344 -- # case "$op" in 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@345 -- # : 1 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # decimal 1 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=1 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 1 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # decimal 2 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=2 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 2 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # return 0 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:09:12.784 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:12.784 --rc genhtml_branch_coverage=1 00:09:12.784 --rc genhtml_function_coverage=1 00:09:12.784 --rc genhtml_legend=1 00:09:12.784 --rc geninfo_all_blocks=1 00:09:12.784 --rc geninfo_unexecuted_blocks=1 00:09:12.784 00:09:12.784 ' 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:09:12.784 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:12.784 --rc genhtml_branch_coverage=1 00:09:12.784 --rc genhtml_function_coverage=1 00:09:12.784 --rc genhtml_legend=1 00:09:12.784 --rc geninfo_all_blocks=1 00:09:12.784 --rc geninfo_unexecuted_blocks=1 00:09:12.784 00:09:12.784 ' 00:09:12.784 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:09:12.784 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:12.784 --rc genhtml_branch_coverage=1 00:09:12.784 --rc genhtml_function_coverage=1 00:09:12.784 --rc genhtml_legend=1 00:09:12.784 --rc geninfo_all_blocks=1 00:09:12.784 --rc geninfo_unexecuted_blocks=1 00:09:12.784 00:09:12.785 ' 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:09:12.785 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:09:12.785 --rc genhtml_branch_coverage=1 00:09:12.785 --rc genhtml_function_coverage=1 00:09:12.785 --rc genhtml_legend=1 00:09:12.785 --rc geninfo_all_blocks=1 00:09:12.785 --rc geninfo_unexecuted_blocks=1 00:09:12.785 00:09:12.785 ' 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # uname -s 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@5 -- # export PATH 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@51 -- # : 0 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:09:12.785 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@22 -- # nvmftestinit 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@474 -- # prepare_net_devs 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@436 -- # local -g is_hw=no 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@438 -- # remove_spdk_ns 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:09:12.785 08:45:02 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # net_devs=() 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # e810=() 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # local -ga e810 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # x722=() 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # local -ga x722 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # mlx=() 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:20.968 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:09:20.969 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:09:20.969 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:09:20.969 Found net devices under 0000:4b:00.0: cvl_0_0 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:09:20.969 Found net devices under 0000:4b:00.1: cvl_0_1 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # is_hw=yes 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:20.969 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:09:20.970 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:20.970 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.507 ms 00:09:20.970 00:09:20.970 --- 10.0.0.2 ping statistics --- 00:09:20.970 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:20.970 rtt min/avg/max/mdev = 0.507/0.507/0.507/0.000 ms 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:20.970 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:20.970 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.263 ms 00:09:20.970 00:09:20.970 --- 10.0.0.1 ping statistics --- 00:09:20.970 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:20.970 rtt min/avg/max/mdev = 0.263/0.263/0.263/0.000 ms 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@448 -- # return 0 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@23 -- # nvmfappstart -m 0xE 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@724 -- # xtrace_disable 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@507 -- # nvmfpid=1679120 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@508 -- # waitforlisten 1679120 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@831 -- # '[' -z 1679120 ']' 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@836 -- # local max_retries=100 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:20.970 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@840 -- # xtrace_disable 00:09:20.970 08:45:09 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:09:20.970 [2024-11-06 08:45:09.986309] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:09:20.970 [2024-11-06 08:45:09.986369] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:20.970 [2024-11-06 08:45:10.088605] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:09:20.970 [2024-11-06 08:45:10.137077] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:09:20.970 [2024-11-06 08:45:10.137126] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:09:20.970 [2024-11-06 08:45:10.137136] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:09:20.970 [2024-11-06 08:45:10.137143] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:09:20.970 [2024-11-06 08:45:10.137149] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:09:20.970 [2024-11-06 08:45:10.138961] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:09:20.970 [2024-11-06 08:45:10.139096] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:09:20.970 [2024-11-06 08:45:10.139096] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:09:20.970 08:45:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:09:20.970 08:45:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@864 -- # return 0 00:09:20.970 08:45:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:09:20.970 08:45:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@730 -- # xtrace_disable 00:09:20.970 08:45:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:09:20.970 08:45:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:09:20.970 08:45:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@25 -- # null_size=1000 00:09:20.970 08:45:10 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:09:20.970 [2024-11-06 08:45:10.993962] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:09:20.970 08:45:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:09:21.231 08:45:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:09:21.492 [2024-11-06 08:45:11.355358] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:09:21.492 08:45:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:09:21.492 08:45:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 512 -b Malloc0 00:09:21.754 Malloc0 00:09:21.754 08:45:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:09:22.014 Delay0 00:09:22.015 08:45:11 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:22.276 08:45:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create NULL1 1000 512 00:09:22.276 NULL1 00:09:22.276 08:45:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:09:22.537 08:45:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@42 -- # PERF_PID=1679841 00:09:22.537 08:45:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 30 -q 128 -w randread -o 512 -Q 1000 00:09:22.537 08:45:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:22.537 08:45:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:22.798 08:45:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:22.798 08:45:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1001 00:09:22.798 08:45:12 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1001 00:09:23.060 true 00:09:23.060 08:45:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:23.060 08:45:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:23.320 08:45:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:23.320 08:45:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1002 00:09:23.320 08:45:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1002 00:09:23.579 true 00:09:23.579 08:45:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:23.579 08:45:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:23.840 08:45:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:24.101 08:45:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1003 00:09:24.101 08:45:13 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1003 00:09:24.101 true 00:09:24.101 08:45:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:24.101 08:45:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:24.432 08:45:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:24.432 08:45:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1004 00:09:24.432 08:45:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1004 00:09:24.722 true 00:09:24.722 08:45:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:24.722 08:45:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:24.983 08:45:14 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:24.983 08:45:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1005 00:09:24.983 08:45:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1005 00:09:25.243 true 00:09:25.243 08:45:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:25.243 08:45:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:25.505 08:45:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:25.505 08:45:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1006 00:09:25.765 08:45:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1006 00:09:25.765 true 00:09:25.765 08:45:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:25.765 08:45:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:26.026 08:45:15 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:26.286 08:45:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1007 00:09:26.286 08:45:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1007 00:09:26.286 true 00:09:26.286 08:45:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:26.286 08:45:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:26.547 08:45:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:26.808 08:45:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1008 00:09:26.808 08:45:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1008 00:09:26.808 true 00:09:26.808 08:45:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:26.808 08:45:16 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:27.069 08:45:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:27.329 08:45:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1009 00:09:27.329 08:45:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1009 00:09:27.329 true 00:09:27.589 08:45:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:27.589 08:45:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:27.589 08:45:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:27.848 08:45:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1010 00:09:27.848 08:45:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1010 00:09:28.108 true 00:09:28.108 08:45:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:28.108 08:45:17 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:28.108 08:45:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:28.369 08:45:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1011 00:09:28.369 08:45:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1011 00:09:28.629 true 00:09:28.629 08:45:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:28.629 08:45:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:28.629 08:45:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:28.889 08:45:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1012 00:09:28.889 08:45:18 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1012 00:09:29.148 true 00:09:29.148 08:45:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:29.148 08:45:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:29.408 08:45:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:29.408 08:45:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1013 00:09:29.408 08:45:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1013 00:09:29.667 true 00:09:29.667 08:45:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:29.668 08:45:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:29.927 08:45:19 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:29.927 08:45:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1014 00:09:29.927 08:45:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1014 00:09:30.187 true 00:09:30.187 08:45:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:30.187 08:45:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:30.447 08:45:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:30.708 08:45:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1015 00:09:30.708 08:45:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1015 00:09:30.708 true 00:09:30.708 08:45:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:30.708 08:45:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:30.969 08:45:20 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:31.229 08:45:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1016 00:09:31.229 08:45:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1016 00:09:31.229 true 00:09:31.229 08:45:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:31.229 08:45:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:31.489 08:45:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:31.750 08:45:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1017 00:09:31.750 08:45:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1017 00:09:31.750 true 00:09:32.010 08:45:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:32.010 08:45:21 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:32.010 08:45:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:32.270 08:45:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1018 00:09:32.270 08:45:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1018 00:09:32.531 true 00:09:32.531 08:45:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:32.531 08:45:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:32.531 08:45:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:32.791 08:45:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1019 00:09:32.791 08:45:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1019 00:09:33.051 true 00:09:33.051 08:45:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:33.051 08:45:22 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:33.312 08:45:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:33.313 08:45:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1020 00:09:33.313 08:45:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1020 00:09:33.574 true 00:09:33.574 08:45:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:33.574 08:45:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:33.834 08:45:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:33.834 08:45:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1021 00:09:33.834 08:45:23 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1021 00:09:34.095 true 00:09:34.095 08:45:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:34.095 08:45:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:34.355 08:45:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:34.356 08:45:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1022 00:09:34.356 08:45:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1022 00:09:34.616 true 00:09:34.616 08:45:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:34.616 08:45:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:34.876 08:45:24 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:35.136 08:45:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1023 00:09:35.136 08:45:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1023 00:09:35.136 true 00:09:35.136 08:45:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:35.136 08:45:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:35.397 08:45:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:35.657 08:45:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1024 00:09:35.657 08:45:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1024 00:09:35.657 true 00:09:35.657 08:45:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:35.657 08:45:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:35.918 08:45:25 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:36.194 08:45:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1025 00:09:36.194 08:45:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1025 00:09:36.194 true 00:09:36.194 08:45:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:36.194 08:45:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:36.455 08:45:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:36.715 08:45:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1026 00:09:36.715 08:45:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1026 00:09:36.715 true 00:09:36.976 08:45:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:36.976 08:45:26 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:36.976 08:45:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:37.236 08:45:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1027 00:09:37.236 08:45:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1027 00:09:37.497 true 00:09:37.497 08:45:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:37.497 08:45:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:37.497 08:45:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:37.757 08:45:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1028 00:09:37.757 08:45:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1028 00:09:38.018 true 00:09:38.018 08:45:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:38.018 08:45:27 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:38.280 08:45:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:38.280 08:45:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1029 00:09:38.280 08:45:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1029 00:09:38.541 true 00:09:38.541 08:45:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:38.541 08:45:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:38.800 08:45:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:38.800 08:45:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1030 00:09:38.800 08:45:28 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1030 00:09:39.060 true 00:09:39.060 08:45:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:39.060 08:45:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:39.319 08:45:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:39.319 08:45:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1031 00:09:39.319 08:45:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1031 00:09:39.578 true 00:09:39.578 08:45:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:39.578 08:45:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:39.839 08:45:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:40.099 08:45:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1032 00:09:40.099 08:45:29 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1032 00:09:40.099 true 00:09:40.099 08:45:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:40.099 08:45:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:40.360 08:45:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:40.622 08:45:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1033 00:09:40.622 08:45:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1033 00:09:40.622 true 00:09:40.622 08:45:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:40.622 08:45:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:40.882 08:45:30 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:41.143 08:45:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1034 00:09:41.143 08:45:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1034 00:09:41.143 true 00:09:41.143 08:45:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:41.143 08:45:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:41.405 08:45:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:41.665 08:45:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1035 00:09:41.665 08:45:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1035 00:09:41.666 true 00:09:41.927 08:45:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:41.927 08:45:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:41.927 08:45:31 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:42.188 08:45:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1036 00:09:42.188 08:45:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1036 00:09:42.188 true 00:09:42.449 08:45:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:42.449 08:45:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:42.449 08:45:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:42.710 08:45:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1037 00:09:42.710 08:45:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1037 00:09:42.971 true 00:09:42.971 08:45:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:42.971 08:45:32 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:42.971 08:45:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:43.233 08:45:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1038 00:09:43.233 08:45:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1038 00:09:43.493 true 00:09:43.493 08:45:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:43.493 08:45:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:43.493 08:45:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:43.754 08:45:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1039 00:09:43.754 08:45:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1039 00:09:44.015 true 00:09:44.015 08:45:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:44.015 08:45:33 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:44.276 08:45:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:44.276 08:45:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1040 00:09:44.276 08:45:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1040 00:09:44.536 true 00:09:44.536 08:45:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:44.536 08:45:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:44.797 08:45:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:44.797 08:45:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1041 00:09:44.797 08:45:34 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1041 00:09:45.059 true 00:09:45.059 08:45:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:45.059 08:45:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:45.320 08:45:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:45.580 08:45:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1042 00:09:45.580 08:45:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1042 00:09:45.580 true 00:09:45.580 08:45:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:45.580 08:45:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:45.841 08:45:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:46.102 08:45:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1043 00:09:46.102 08:45:35 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1043 00:09:46.102 true 00:09:46.102 08:45:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:46.102 08:45:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:46.362 08:45:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:46.624 08:45:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1044 00:09:46.624 08:45:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1044 00:09:46.624 true 00:09:46.624 08:45:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:46.624 08:45:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:46.885 08:45:36 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:47.146 08:45:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1045 00:09:47.146 08:45:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1045 00:09:47.408 true 00:09:47.408 08:45:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:47.408 08:45:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:47.408 08:45:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:47.669 08:45:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1046 00:09:47.669 08:45:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1046 00:09:47.929 true 00:09:47.929 08:45:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:47.929 08:45:37 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:47.929 08:45:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:48.226 08:45:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1047 00:09:48.226 08:45:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1047 00:09:48.514 true 00:09:48.514 08:45:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:48.514 08:45:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:48.514 08:45:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:48.775 08:45:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1048 00:09:48.775 08:45:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1048 00:09:49.036 true 00:09:49.036 08:45:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:49.036 08:45:38 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:49.297 08:45:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:49.297 08:45:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1049 00:09:49.297 08:45:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1049 00:09:49.557 true 00:09:49.557 08:45:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:49.558 08:45:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:49.817 08:45:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:49.818 08:45:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1050 00:09:49.818 08:45:39 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1050 00:09:50.077 true 00:09:50.077 08:45:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:50.077 08:45:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:50.337 08:45:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:50.337 08:45:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1051 00:09:50.337 08:45:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1051 00:09:50.598 true 00:09:50.598 08:45:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:50.598 08:45:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:50.859 08:45:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:51.119 08:45:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1052 00:09:51.119 08:45:40 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1052 00:09:51.119 true 00:09:51.119 08:45:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:51.119 08:45:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:51.379 08:45:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:51.639 08:45:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1053 00:09:51.639 08:45:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1053 00:09:51.639 true 00:09:51.639 08:45:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:51.639 08:45:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:51.899 08:45:41 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:52.159 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1054 00:09:52.159 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1054 00:09:52.159 true 00:09:52.420 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:52.420 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:52.420 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:52.681 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1055 00:09:52.681 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1055 00:09:52.681 Initializing NVMe Controllers 00:09:52.681 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:09:52.681 Controller IO queue size 128, less than required. 00:09:52.681 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:09:52.681 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:09:52.681 Initialization complete. Launching workers. 00:09:52.681 ======================================================== 00:09:52.681 Latency(us) 00:09:52.681 Device Information : IOPS MiB/s Average min max 00:09:52.681 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 30112.03 14.70 4250.61 1423.94 9967.98 00:09:52.681 ======================================================== 00:09:52.681 Total : 30112.03 14.70 4250.61 1423.94 9967.98 00:09:52.681 00:09:52.681 true 00:09:52.941 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 1679841 00:09:52.941 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh: line 44: kill: (1679841) - No such process 00:09:52.941 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@53 -- # wait 1679841 00:09:52.941 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:52.941 08:45:42 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:53.201 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # nthreads=8 00:09:53.201 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # pids=() 00:09:53.201 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i = 0 )) 00:09:53.201 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:53.201 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null0 100 4096 00:09:53.461 null0 00:09:53.461 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:53.461 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:53.461 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null1 100 4096 00:09:53.461 null1 00:09:53.461 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:53.461 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:53.461 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null2 100 4096 00:09:53.721 null2 00:09:53.721 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:53.721 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:53.721 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null3 100 4096 00:09:53.982 null3 00:09:53.982 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:53.982 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:53.982 08:45:43 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null4 100 4096 00:09:53.982 null4 00:09:53.982 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:53.982 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:53.982 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null5 100 4096 00:09:54.242 null5 00:09:54.242 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:54.242 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:54.242 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null6 100 4096 00:09:54.503 null6 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null7 100 4096 00:09:54.503 null7 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i = 0 )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 1 null0 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=1 bdev=null0 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 2 null1 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=2 bdev=null1 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:54.503 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 3 null2 00:09:54.763 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=3 bdev=null2 00:09:54.763 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 4 null3 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=4 bdev=null3 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 5 null4 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=5 bdev=null4 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 6 null5 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=6 bdev=null5 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 7 null6 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=7 bdev=null6 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@66 -- # wait 1686397 1686398 1686400 1686402 1686404 1686406 1686408 1686409 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 8 null7 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=8 bdev=null7 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:54.764 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:55.027 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.027 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.027 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:55.027 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.027 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.027 08:45:44 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.027 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.289 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:55.551 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:55.552 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:55.814 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:56.076 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:56.076 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:56.076 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:56.076 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:56.076 08:45:45 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.076 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.338 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:56.600 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:56.862 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:57.124 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:57.124 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:57.124 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.124 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.124 08:45:46 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.124 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:57.385 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:57.645 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.646 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.646 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:57.646 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.646 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.646 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:57.646 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.646 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.646 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:57.646 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.907 08:45:47 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:57.907 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.907 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.907 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:57.907 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:57.907 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:57.907 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.169 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@70 -- # nvmftestfini 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@514 -- # nvmfcleanup 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@121 -- # sync 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@124 -- # set +e 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:09:58.429 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:09:58.429 rmmod nvme_tcp 00:09:58.429 rmmod nvme_fabrics 00:09:58.429 rmmod nvme_keyring 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@128 -- # set -e 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@129 -- # return 0 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@515 -- # '[' -n 1679120 ']' 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@516 -- # killprocess 1679120 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@950 -- # '[' -z 1679120 ']' 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@954 -- # kill -0 1679120 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@955 -- # uname 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1679120 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1679120' 00:09:58.690 killing process with pid 1679120 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@969 -- # kill 1679120 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@974 -- # wait 1679120 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@297 -- # iptr 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # iptables-save 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # iptables-restore 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:09:58.690 08:45:48 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:01.236 08:45:50 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:10:01.236 00:10:01.236 real 0m48.263s 00:10:01.236 user 3m18.682s 00:10:01.236 sys 0m16.613s 00:10:01.236 08:45:50 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:01.236 08:45:50 nvmf_tcp.nvmf_target_core.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:10:01.236 ************************************ 00:10:01.236 END TEST nvmf_ns_hotplug_stress 00:10:01.236 ************************************ 00:10:01.237 08:45:50 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@23 -- # run_test nvmf_delete_subsystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:10:01.237 08:45:50 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:10:01.237 08:45:50 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:01.237 08:45:50 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:10:01.237 ************************************ 00:10:01.237 START TEST nvmf_delete_subsystem 00:10:01.237 ************************************ 00:10:01.237 08:45:50 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:10:01.237 * Looking for test storage... 00:10:01.237 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:01.237 08:45:50 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1689 -- # lcov --version 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@336 -- # IFS=.-: 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@336 -- # read -ra ver1 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@337 -- # IFS=.-: 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@337 -- # read -ra ver2 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@338 -- # local 'op=<' 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@340 -- # ver1_l=2 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@341 -- # ver2_l=1 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@344 -- # case "$op" in 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@345 -- # : 1 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@365 -- # decimal 1 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=1 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 1 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@365 -- # ver1[v]=1 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@366 -- # decimal 2 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=2 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 2 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@366 -- # ver2[v]=2 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@368 -- # return 0 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:10:01.237 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:01.237 --rc genhtml_branch_coverage=1 00:10:01.237 --rc genhtml_function_coverage=1 00:10:01.237 --rc genhtml_legend=1 00:10:01.237 --rc geninfo_all_blocks=1 00:10:01.237 --rc geninfo_unexecuted_blocks=1 00:10:01.237 00:10:01.237 ' 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:10:01.237 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:01.237 --rc genhtml_branch_coverage=1 00:10:01.237 --rc genhtml_function_coverage=1 00:10:01.237 --rc genhtml_legend=1 00:10:01.237 --rc geninfo_all_blocks=1 00:10:01.237 --rc geninfo_unexecuted_blocks=1 00:10:01.237 00:10:01.237 ' 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:10:01.237 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:01.237 --rc genhtml_branch_coverage=1 00:10:01.237 --rc genhtml_function_coverage=1 00:10:01.237 --rc genhtml_legend=1 00:10:01.237 --rc geninfo_all_blocks=1 00:10:01.237 --rc geninfo_unexecuted_blocks=1 00:10:01.237 00:10:01.237 ' 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:10:01.237 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:01.237 --rc genhtml_branch_coverage=1 00:10:01.237 --rc genhtml_function_coverage=1 00:10:01.237 --rc genhtml_legend=1 00:10:01.237 --rc geninfo_all_blocks=1 00:10:01.237 --rc geninfo_unexecuted_blocks=1 00:10:01.237 00:10:01.237 ' 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # uname -s 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@15 -- # shopt -s extglob 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:01.237 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@5 -- # export PATH 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@51 -- # : 0 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:01.238 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@12 -- # nvmftestinit 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@474 -- # prepare_net_devs 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@436 -- # local -g is_hw=no 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@438 -- # remove_spdk_ns 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@309 -- # xtrace_disable 00:10:01.238 08:45:51 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # pci_devs=() 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # net_devs=() 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # e810=() 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # local -ga e810 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # x722=() 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # local -ga x722 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # mlx=() 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # local -ga mlx 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:10:09.386 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:09.386 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:10:09.386 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:10:09.387 Found net devices under 0000:4b:00.0: cvl_0_0 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:10:09.387 Found net devices under 0000:4b:00.1: cvl_0_1 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # is_hw=yes 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:10:09.387 08:45:57 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:10:09.387 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:09.387 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.695 ms 00:10:09.387 00:10:09.387 --- 10.0.0.2 ping statistics --- 00:10:09.387 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:09.387 rtt min/avg/max/mdev = 0.695/0.695/0.695/0.000 ms 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:09.387 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:09.387 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.311 ms 00:10:09.387 00:10:09.387 --- 10.0.0.1 ping statistics --- 00:10:09.387 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:09.387 rtt min/avg/max/mdev = 0.311/0.311/0.311/0.000 ms 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@448 -- # return 0 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@13 -- # nvmfappstart -m 0x3 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@724 -- # xtrace_disable 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@507 -- # nvmfpid=1691576 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@508 -- # waitforlisten 1691576 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@831 -- # '[' -z 1691576 ']' 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:09.387 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:09.387 08:45:58 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:09.387 [2024-11-06 08:45:58.349210] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:10:09.387 [2024-11-06 08:45:58.349275] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:09.387 [2024-11-06 08:45:58.432165] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:10:09.387 [2024-11-06 08:45:58.473683] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:09.387 [2024-11-06 08:45:58.473720] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:09.387 [2024-11-06 08:45:58.473727] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:09.387 [2024-11-06 08:45:58.473734] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:09.387 [2024-11-06 08:45:58.473740] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:09.387 [2024-11-06 08:45:58.475056] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:10:09.387 [2024-11-06 08:45:58.475059] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:09.387 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:09.387 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@864 -- # return 0 00:10:09.387 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:10:09.387 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@730 -- # xtrace_disable 00:10:09.387 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:09.387 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:09.387 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:09.388 [2024-11-06 08:45:59.190431] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:09.388 [2024-11-06 08:45:59.214613] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:09.388 NULL1 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@23 -- # rpc_cmd bdev_delay_create -b NULL1 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:09.388 Delay0 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@28 -- # perf_pid=1691927 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@30 -- # sleep 2 00:10:09.388 08:45:59 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 5 -q 128 -w randrw -M 70 -o 512 -P 4 00:10:09.388 [2024-11-06 08:45:59.311404] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:10:11.310 08:46:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@32 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:11.310 08:46:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:11.310 08:46:01 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 starting I/O failed: -6 00:10:11.572 [2024-11-06 08:46:01.434568] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15b5960 is same with the state(6) to be set 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.572 Write completed with error (sct=0, sc=8) 00:10:11.572 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 starting I/O failed: -6 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 starting I/O failed: -6 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 starting I/O failed: -6 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 starting I/O failed: -6 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 starting I/O failed: -6 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 starting I/O failed: -6 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 starting I/O failed: -6 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 starting I/O failed: -6 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 starting I/O failed: -6 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 starting I/O failed: -6 00:10:11.573 starting I/O failed: -6 00:10:11.573 starting I/O failed: -6 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Read completed with error (sct=0, sc=8) 00:10:11.573 Write completed with error (sct=0, sc=8) 00:10:11.573 [2024-11-06 08:46:01.440116] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f4b1c00d450 is same with the state(6) to be set 00:10:12.516 [2024-11-06 08:46:02.408918] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15b69a0 is same with the state(6) to be set 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 [2024-11-06 08:46:02.437805] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15b5b40 is same with the state(6) to be set 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 [2024-11-06 08:46:02.438282] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15b5780 is same with the state(6) to be set 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.516 Read completed with error (sct=0, sc=8) 00:10:12.516 Write completed with error (sct=0, sc=8) 00:10:12.517 Write completed with error (sct=0, sc=8) 00:10:12.517 Write completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 [2024-11-06 08:46:02.443176] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f4b1c00cfe0 is same with the state(6) to be set 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Write completed with error (sct=0, sc=8) 00:10:12.517 Write completed with error (sct=0, sc=8) 00:10:12.517 Write completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Write completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Write completed with error (sct=0, sc=8) 00:10:12.517 Read completed with error (sct=0, sc=8) 00:10:12.517 Write completed with error (sct=0, sc=8) 00:10:12.517 [2024-11-06 08:46:02.443432] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f4b1c00d780 is same with the state(6) to be set 00:10:12.517 Initializing NVMe Controllers 00:10:12.517 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:10:12.517 Controller IO queue size 128, less than required. 00:10:12.517 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:10:12.517 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:10:12.517 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:10:12.517 Initialization complete. Launching workers. 00:10:12.517 ======================================================== 00:10:12.517 Latency(us) 00:10:12.517 Device Information : IOPS MiB/s Average min max 00:10:12.517 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 170.89 0.08 892004.59 246.79 1005776.44 00:10:12.517 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 156.44 0.08 961148.13 369.78 2003522.76 00:10:12.517 ======================================================== 00:10:12.517 Total : 327.33 0.16 925050.36 246.79 2003522.76 00:10:12.517 00:10:12.517 [2024-11-06 08:46:02.444112] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x15b69a0 (9): Bad file descriptor 00:10:12.517 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:10:12.517 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:12.517 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@34 -- # delay=0 00:10:12.517 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 1691927 00:10:12.517 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@36 -- # sleep 0.5 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@38 -- # (( delay++ > 30 )) 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 1691927 00:10:13.087 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 35: kill: (1691927) - No such process 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@45 -- # NOT wait 1691927 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@650 -- # local es=0 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@652 -- # valid_exec_arg wait 1691927 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@638 -- # local arg=wait 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # type -t wait 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@653 -- # wait 1691927 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@653 -- # es=1 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:10:13.087 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:13.088 [2024-11-06 08:46:02.974674] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@50 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@54 -- # perf_pid=1692609 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@56 -- # delay=0 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 1692609 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 3 -q 128 -w randrw -M 70 -o 512 -P 4 00:10:13.088 08:46:02 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:13.088 [2024-11-06 08:46:03.051621] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:10:13.658 08:46:03 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:13.658 08:46:03 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 1692609 00:10:13.658 08:46:03 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:13.918 08:46:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:13.918 08:46:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 1692609 00:10:13.918 08:46:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:14.488 08:46:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:14.488 08:46:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 1692609 00:10:14.488 08:46:04 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:15.060 08:46:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:15.060 08:46:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 1692609 00:10:15.060 08:46:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:15.630 08:46:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:15.630 08:46:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 1692609 00:10:15.630 08:46:05 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:16.200 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:16.200 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 1692609 00:10:16.200 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:16.200 [2024-11-06 08:46:06.261879] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2014d50 is same with the state(6) to be set 00:10:16.200 Initializing NVMe Controllers 00:10:16.200 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:10:16.200 Controller IO queue size 128, less than required. 00:10:16.200 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:10:16.200 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:10:16.200 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:10:16.200 Initialization complete. Launching workers. 00:10:16.200 ======================================================== 00:10:16.200 Latency(us) 00:10:16.200 Device Information : IOPS MiB/s Average min max 00:10:16.200 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 128.00 0.06 1002075.41 1000099.20 1041685.01 00:10:16.200 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 128.00 0.06 1002893.10 1000185.57 1009496.98 00:10:16.200 ======================================================== 00:10:16.200 Total : 256.00 0.12 1002484.25 1000099.20 1041685.01 00:10:16.200 00:10:16.200 [2024-11-06 08:46:06.262501] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20a3a80 is same with the state(6) to be set 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 1692609 00:10:16.460 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 57: kill: (1692609) - No such process 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@67 -- # wait 1692609 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- target/delete_subsystem.sh@71 -- # nvmftestfini 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@514 -- # nvmfcleanup 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@121 -- # sync 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@124 -- # set +e 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:10:16.460 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:10:16.460 rmmod nvme_tcp 00:10:16.460 rmmod nvme_fabrics 00:10:16.460 rmmod nvme_keyring 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@128 -- # set -e 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@129 -- # return 0 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@515 -- # '[' -n 1691576 ']' 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@516 -- # killprocess 1691576 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@950 -- # '[' -z 1691576 ']' 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@954 -- # kill -0 1691576 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@955 -- # uname 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1691576 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1691576' 00:10:16.720 killing process with pid 1691576 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@969 -- # kill 1691576 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@974 -- # wait 1691576 00:10:16.720 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@297 -- # iptr 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # iptables-save 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # iptables-restore 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:16.721 08:46:06 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:19.266 08:46:08 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:10:19.266 00:10:19.266 real 0m17.967s 00:10:19.266 user 0m30.625s 00:10:19.266 sys 0m6.491s 00:10:19.266 08:46:08 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:19.266 08:46:08 nvmf_tcp.nvmf_target_core.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:19.266 ************************************ 00:10:19.266 END TEST nvmf_delete_subsystem 00:10:19.266 ************************************ 00:10:19.266 08:46:08 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@26 -- # run_test nvmf_host_management /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:10:19.266 08:46:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:10:19.266 08:46:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:19.266 08:46:08 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:10:19.266 ************************************ 00:10:19.266 START TEST nvmf_host_management 00:10:19.266 ************************************ 00:10:19.266 08:46:08 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:10:19.266 * Looking for test storage... 00:10:19.266 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1689 -- # lcov --version 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@336 -- # IFS=.-: 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@336 -- # read -ra ver1 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@337 -- # IFS=.-: 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@337 -- # read -ra ver2 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@338 -- # local 'op=<' 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@340 -- # ver1_l=2 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@341 -- # ver2_l=1 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@344 -- # case "$op" in 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@345 -- # : 1 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@365 -- # decimal 1 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@353 -- # local d=1 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@355 -- # echo 1 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@365 -- # ver1[v]=1 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@366 -- # decimal 2 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@353 -- # local d=2 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@355 -- # echo 2 00:10:19.266 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@366 -- # ver2[v]=2 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@368 -- # return 0 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:10:19.267 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:19.267 --rc genhtml_branch_coverage=1 00:10:19.267 --rc genhtml_function_coverage=1 00:10:19.267 --rc genhtml_legend=1 00:10:19.267 --rc geninfo_all_blocks=1 00:10:19.267 --rc geninfo_unexecuted_blocks=1 00:10:19.267 00:10:19.267 ' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:10:19.267 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:19.267 --rc genhtml_branch_coverage=1 00:10:19.267 --rc genhtml_function_coverage=1 00:10:19.267 --rc genhtml_legend=1 00:10:19.267 --rc geninfo_all_blocks=1 00:10:19.267 --rc geninfo_unexecuted_blocks=1 00:10:19.267 00:10:19.267 ' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:10:19.267 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:19.267 --rc genhtml_branch_coverage=1 00:10:19.267 --rc genhtml_function_coverage=1 00:10:19.267 --rc genhtml_legend=1 00:10:19.267 --rc geninfo_all_blocks=1 00:10:19.267 --rc geninfo_unexecuted_blocks=1 00:10:19.267 00:10:19.267 ' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:10:19.267 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:19.267 --rc genhtml_branch_coverage=1 00:10:19.267 --rc genhtml_function_coverage=1 00:10:19.267 --rc genhtml_legend=1 00:10:19.267 --rc geninfo_all_blocks=1 00:10:19.267 --rc geninfo_unexecuted_blocks=1 00:10:19.267 00:10:19.267 ' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@7 -- # uname -s 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@15 -- # shopt -s extglob 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@5 -- # export PATH 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@51 -- # : 0 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:19.267 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@11 -- # MALLOC_BDEV_SIZE=64 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@105 -- # nvmftestinit 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@474 -- # prepare_net_devs 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@436 -- # local -g is_hw=no 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@438 -- # remove_spdk_ns 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@309 -- # xtrace_disable 00:10:19.267 08:46:09 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@315 -- # pci_devs=() 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@315 -- # local -a pci_devs 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@316 -- # pci_net_devs=() 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@317 -- # pci_drivers=() 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@317 -- # local -A pci_drivers 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@319 -- # net_devs=() 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@319 -- # local -ga net_devs 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@320 -- # e810=() 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@320 -- # local -ga e810 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@321 -- # x722=() 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@321 -- # local -ga x722 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@322 -- # mlx=() 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@322 -- # local -ga mlx 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:27.411 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:10:27.412 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:10:27.412 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@416 -- # [[ up == up ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:10:27.412 Found net devices under 0000:4b:00.0: cvl_0_0 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@416 -- # [[ up == up ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:10:27.412 Found net devices under 0000:4b:00.1: cvl_0_1 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@440 -- # is_hw=yes 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:10:27.412 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:27.412 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.690 ms 00:10:27.412 00:10:27.412 --- 10.0.0.2 ping statistics --- 00:10:27.412 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:27.412 rtt min/avg/max/mdev = 0.690/0.690/0.690/0.000 ms 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:27.412 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:27.412 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.281 ms 00:10:27.412 00:10:27.412 --- 10.0.0.1 ping statistics --- 00:10:27.412 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:27.412 rtt min/avg/max/mdev = 0.281/0.281/0.281/0.000 ms 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@448 -- # return 0 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@107 -- # nvmf_host_management 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@69 -- # starttarget 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@16 -- # nvmfappstart -m 0x1E 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@724 -- # xtrace_disable 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@507 -- # nvmfpid=1697628 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@508 -- # waitforlisten 1697628 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@831 -- # '[' -z 1697628 ']' 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:27.412 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:27.413 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:27.413 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:27.413 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:27.413 08:46:16 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:27.413 [2024-11-06 08:46:16.623531] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:10:27.413 [2024-11-06 08:46:16.623595] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:27.413 [2024-11-06 08:46:16.722199] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:10:27.413 [2024-11-06 08:46:16.774862] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:27.413 [2024-11-06 08:46:16.774913] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:27.413 [2024-11-06 08:46:16.774922] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:27.413 [2024-11-06 08:46:16.774930] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:27.413 [2024-11-06 08:46:16.774936] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:27.413 [2024-11-06 08:46:16.777268] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:10:27.413 [2024-11-06 08:46:16.777440] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:10:27.413 [2024-11-06 08:46:16.777605] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:10:27.413 [2024-11-06 08:46:16.777605] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@864 -- # return 0 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@730 -- # xtrace_disable 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:27.413 [2024-11-06 08:46:17.482250] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@20 -- # timing_enter create_subsystem 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@724 -- # xtrace_disable 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@22 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@23 -- # cat 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@30 -- # rpc_cmd 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:27.413 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:27.673 Malloc0 00:10:27.673 [2024-11-06 08:46:17.561938] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@31 -- # timing_exit create_subsystems 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@730 -- # xtrace_disable 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@73 -- # perfpid=1697822 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@74 -- # waitforlisten 1697822 /var/tmp/bdevperf.sock 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@831 -- # '[' -z 1697822 ']' 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:10:27.673 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@72 -- # gen_nvmf_target_json 0 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@558 -- # config=() 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@558 -- # local subsystem config 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:10:27.673 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:10:27.673 { 00:10:27.673 "params": { 00:10:27.673 "name": "Nvme$subsystem", 00:10:27.673 "trtype": "$TEST_TRANSPORT", 00:10:27.673 "traddr": "$NVMF_FIRST_TARGET_IP", 00:10:27.673 "adrfam": "ipv4", 00:10:27.673 "trsvcid": "$NVMF_PORT", 00:10:27.673 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:10:27.674 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:10:27.674 "hdgst": ${hdgst:-false}, 00:10:27.674 "ddgst": ${ddgst:-false} 00:10:27.674 }, 00:10:27.674 "method": "bdev_nvme_attach_controller" 00:10:27.674 } 00:10:27.674 EOF 00:10:27.674 )") 00:10:27.674 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@580 -- # cat 00:10:27.674 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # jq . 00:10:27.674 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@583 -- # IFS=, 00:10:27.674 08:46:17 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:10:27.674 "params": { 00:10:27.674 "name": "Nvme0", 00:10:27.674 "trtype": "tcp", 00:10:27.674 "traddr": "10.0.0.2", 00:10:27.674 "adrfam": "ipv4", 00:10:27.674 "trsvcid": "4420", 00:10:27.674 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:10:27.674 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:10:27.674 "hdgst": false, 00:10:27.674 "ddgst": false 00:10:27.674 }, 00:10:27.674 "method": "bdev_nvme_attach_controller" 00:10:27.674 }' 00:10:27.674 [2024-11-06 08:46:17.673895] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:10:27.674 [2024-11-06 08:46:17.673950] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1697822 ] 00:10:27.674 [2024-11-06 08:46:17.745317] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:27.674 [2024-11-06 08:46:17.781449] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:27.934 Running I/O for 10 seconds... 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@864 -- # return 0 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@75 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@78 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@80 -- # waitforio /var/tmp/bdevperf.sock Nvme0n1 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@45 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@49 -- # '[' -z Nvme0n1 ']' 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@52 -- # local ret=1 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@53 -- # local i 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@54 -- # (( i = 10 )) 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=771 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@58 -- # '[' 771 -ge 100 ']' 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@59 -- # ret=0 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@60 -- # break 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@64 -- # return 0 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@84 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.506 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:28.507 [2024-11-06 08:46:18.546140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:114560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:106496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:106624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:106752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:106880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:107008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:107136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:107264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:107392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:107520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:107648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546383] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:107776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:107904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546408] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:108032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:108160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:108288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:108416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:108544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:108672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:108800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:108928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:109056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:109184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:109312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:109440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:109568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:109696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:109824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:109952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:110080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:110208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:110336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:110464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:110592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:110720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:110848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:110976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:111104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:111232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.507 [2024-11-06 08:46:18.546873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:111360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.507 [2024-11-06 08:46:18.546880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.546890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:111488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.546898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.546907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:111616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.546915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.546924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:111744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.546931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.546941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:111872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.546949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.546958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:112000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.546966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.546975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:112128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.546982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.546992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:112256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:112384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:112512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:112640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:112768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:112896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:113024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:113152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:113280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:113408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:113536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:113664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:113792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:113920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:114048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:114176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:114304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:114432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:10:28.508 [2024-11-06 08:46:18.547298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.547307] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10edff0 is same with the state(6) to be set 00:10:28.508 [2024-11-06 08:46:18.548587] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:10:28.508 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.508 task offset: 114560 on job bdev=Nvme0n1 fails 00:10:28.508 00:10:28.508 Latency(us) 00:10:28.508 [2024-11-06T07:46:18.621Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:28.508 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:10:28.508 Job: Nvme0n1 ended in about 0.58 seconds with error 00:10:28.508 Verification LBA range: start 0x0 length 0x400 00:10:28.508 Nvme0n1 : 0.58 1441.28 90.08 110.87 0.00 40243.34 5188.27 34078.72 00:10:28.508 [2024-11-06T07:46:18.621Z] =================================================================================================================== 00:10:28.508 [2024-11-06T07:46:18.621Z] Total : 1441.28 90.08 110.87 0.00 40243.34 5188.27 34078.72 00:10:28.508 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@85 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:10:28.508 [2024-11-06 08:46:18.550599] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:10:28.508 [2024-11-06 08:46:18.550623] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xed5000 (9): Bad file descriptor 00:10:28.508 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:10:28.508 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:28.508 [2024-11-06 08:46:18.554934] ctrlr.c: 823:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode0' does not allow host 'nqn.2016-06.io.spdk:host0' 00:10:28.508 [2024-11-06 08:46:18.555031] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:3 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:10:28.508 [2024-11-06 08:46:18.555054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND SPECIFIC (01/84) qid:0 cid:3 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:10:28.508 [2024-11-06 08:46:18.555070] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode0 00:10:28.508 [2024-11-06 08:46:18.555078] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 132 00:10:28.508 [2024-11-06 08:46:18.555086] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:10:28.508 [2024-11-06 08:46:18.555093] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0xed5000 00:10:28.508 [2024-11-06 08:46:18.555120] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xed5000 (9): Bad file descriptor 00:10:28.508 [2024-11-06 08:46:18.555133] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:10:28.508 [2024-11-06 08:46:18.555140] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:10:28.508 [2024-11-06 08:46:18.555149] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:10:28.508 [2024-11-06 08:46:18.555164] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:10:28.508 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:10:28.508 08:46:18 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@87 -- # sleep 1 00:10:29.892 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@91 -- # kill -9 1697822 00:10:29.892 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh: line 91: kill: (1697822) - No such process 00:10:29.892 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@91 -- # true 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@97 -- # rm -f /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 /var/tmp/spdk_cpu_lock_003 /var/tmp/spdk_cpu_lock_004 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@100 -- # gen_nvmf_target_json 0 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@558 -- # config=() 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@558 -- # local subsystem config 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:10:29.893 { 00:10:29.893 "params": { 00:10:29.893 "name": "Nvme$subsystem", 00:10:29.893 "trtype": "$TEST_TRANSPORT", 00:10:29.893 "traddr": "$NVMF_FIRST_TARGET_IP", 00:10:29.893 "adrfam": "ipv4", 00:10:29.893 "trsvcid": "$NVMF_PORT", 00:10:29.893 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:10:29.893 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:10:29.893 "hdgst": ${hdgst:-false}, 00:10:29.893 "ddgst": ${ddgst:-false} 00:10:29.893 }, 00:10:29.893 "method": "bdev_nvme_attach_controller" 00:10:29.893 } 00:10:29.893 EOF 00:10:29.893 )") 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@580 -- # cat 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@582 -- # jq . 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@583 -- # IFS=, 00:10:29.893 08:46:19 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:10:29.893 "params": { 00:10:29.893 "name": "Nvme0", 00:10:29.893 "trtype": "tcp", 00:10:29.893 "traddr": "10.0.0.2", 00:10:29.893 "adrfam": "ipv4", 00:10:29.893 "trsvcid": "4420", 00:10:29.893 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:10:29.893 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:10:29.893 "hdgst": false, 00:10:29.893 "ddgst": false 00:10:29.893 }, 00:10:29.893 "method": "bdev_nvme_attach_controller" 00:10:29.893 }' 00:10:29.893 [2024-11-06 08:46:19.630010] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:10:29.893 [2024-11-06 08:46:19.630072] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1698308 ] 00:10:29.893 [2024-11-06 08:46:19.699888] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:29.893 [2024-11-06 08:46:19.735601] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:29.893 Running I/O for 1 seconds... 00:10:31.162 1600.00 IOPS, 100.00 MiB/s 00:10:31.162 Latency(us) 00:10:31.162 [2024-11-06T07:46:21.275Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:10:31.162 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:10:31.162 Verification LBA range: start 0x0 length 0x400 00:10:31.162 Nvme0n1 : 1.05 1581.43 98.84 0.00 0.00 38290.72 5761.71 52647.25 00:10:31.162 [2024-11-06T07:46:21.275Z] =================================================================================================================== 00:10:31.162 [2024-11-06T07:46:21.275Z] Total : 1581.43 98.84 0.00 0.00 38290.72 5761.71 52647.25 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@102 -- # stoptarget 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@36 -- # rm -f ./local-job0-0-verify.state 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@37 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@38 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@40 -- # nvmftestfini 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@514 -- # nvmfcleanup 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@121 -- # sync 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@124 -- # set +e 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@125 -- # for i in {1..20} 00:10:31.162 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:10:31.162 rmmod nvme_tcp 00:10:31.162 rmmod nvme_fabrics 00:10:31.163 rmmod nvme_keyring 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@128 -- # set -e 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@129 -- # return 0 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@515 -- # '[' -n 1697628 ']' 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@516 -- # killprocess 1697628 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@950 -- # '[' -z 1697628 ']' 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@954 -- # kill -0 1697628 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@955 -- # uname 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1697628 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1697628' 00:10:31.163 killing process with pid 1697628 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@969 -- # kill 1697628 00:10:31.163 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@974 -- # wait 1697628 00:10:31.501 [2024-11-06 08:46:21.333123] app.c: 721:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 1, errno: 2 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@297 -- # iptr 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@789 -- # iptables-save 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@789 -- # iptables-restore 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@302 -- # remove_spdk_ns 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:31.501 08:46:21 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:33.428 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_host_management -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:10:33.428 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_host_management -- target/host_management.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:10:33.428 00:10:33.428 real 0m14.489s 00:10:33.428 user 0m23.012s 00:10:33.428 sys 0m6.558s 00:10:33.428 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:33.428 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:10:33.428 ************************************ 00:10:33.428 END TEST nvmf_host_management 00:10:33.428 ************************************ 00:10:33.428 08:46:23 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@27 -- # run_test nvmf_lvol /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:10:33.428 08:46:23 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:10:33.428 08:46:23 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:33.428 08:46:23 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:10:33.428 ************************************ 00:10:33.428 START TEST nvmf_lvol 00:10:33.429 ************************************ 00:10:33.429 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:10:33.691 * Looking for test storage... 00:10:33.691 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1689 -- # lcov --version 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@336 -- # IFS=.-: 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@336 -- # read -ra ver1 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@337 -- # IFS=.-: 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@337 -- # read -ra ver2 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@338 -- # local 'op=<' 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@340 -- # ver1_l=2 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@341 -- # ver2_l=1 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@344 -- # case "$op" in 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@345 -- # : 1 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@365 -- # decimal 1 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@353 -- # local d=1 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@355 -- # echo 1 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@365 -- # ver1[v]=1 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@366 -- # decimal 2 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@353 -- # local d=2 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@355 -- # echo 2 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@366 -- # ver2[v]=2 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@368 -- # return 0 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:10:33.691 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:33.691 --rc genhtml_branch_coverage=1 00:10:33.691 --rc genhtml_function_coverage=1 00:10:33.691 --rc genhtml_legend=1 00:10:33.691 --rc geninfo_all_blocks=1 00:10:33.691 --rc geninfo_unexecuted_blocks=1 00:10:33.691 00:10:33.691 ' 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:10:33.691 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:33.691 --rc genhtml_branch_coverage=1 00:10:33.691 --rc genhtml_function_coverage=1 00:10:33.691 --rc genhtml_legend=1 00:10:33.691 --rc geninfo_all_blocks=1 00:10:33.691 --rc geninfo_unexecuted_blocks=1 00:10:33.691 00:10:33.691 ' 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:10:33.691 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:33.691 --rc genhtml_branch_coverage=1 00:10:33.691 --rc genhtml_function_coverage=1 00:10:33.691 --rc genhtml_legend=1 00:10:33.691 --rc geninfo_all_blocks=1 00:10:33.691 --rc geninfo_unexecuted_blocks=1 00:10:33.691 00:10:33.691 ' 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:10:33.691 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:33.691 --rc genhtml_branch_coverage=1 00:10:33.691 --rc genhtml_function_coverage=1 00:10:33.691 --rc genhtml_legend=1 00:10:33.691 --rc geninfo_all_blocks=1 00:10:33.691 --rc geninfo_unexecuted_blocks=1 00:10:33.691 00:10:33.691 ' 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@7 -- # uname -s 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@15 -- # shopt -s extglob 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:33.691 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@5 -- # export PATH 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@51 -- # : 0 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:33.692 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@11 -- # MALLOC_BDEV_SIZE=64 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@13 -- # LVOL_BDEV_INIT_SIZE=20 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@14 -- # LVOL_BDEV_FINAL_SIZE=30 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@18 -- # nvmftestinit 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@474 -- # prepare_net_devs 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@436 -- # local -g is_hw=no 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@438 -- # remove_spdk_ns 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@309 -- # xtrace_disable 00:10:33.692 08:46:23 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@315 -- # pci_devs=() 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@315 -- # local -a pci_devs 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@316 -- # pci_net_devs=() 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@317 -- # pci_drivers=() 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@317 -- # local -A pci_drivers 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@319 -- # net_devs=() 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@319 -- # local -ga net_devs 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@320 -- # e810=() 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@320 -- # local -ga e810 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@321 -- # x722=() 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@321 -- # local -ga x722 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@322 -- # mlx=() 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@322 -- # local -ga mlx 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:10:40.283 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:10:40.283 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@416 -- # [[ up == up ]] 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:40.283 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:10:40.283 Found net devices under 0000:4b:00.0: cvl_0_0 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@416 -- # [[ up == up ]] 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:10:40.284 Found net devices under 0000:4b:00.1: cvl_0_1 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@440 -- # is_hw=yes 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:10:40.284 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:40.544 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:40.544 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:40.544 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:10:40.544 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:40.544 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:40.544 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:40.545 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:10:40.545 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:10:40.545 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:40.545 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.584 ms 00:10:40.545 00:10:40.545 --- 10.0.0.2 ping statistics --- 00:10:40.545 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:40.545 rtt min/avg/max/mdev = 0.584/0.584/0.584/0.000 ms 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:40.806 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:40.806 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.356 ms 00:10:40.806 00:10:40.806 --- 10.0.0.1 ping statistics --- 00:10:40.806 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:40.806 rtt min/avg/max/mdev = 0.356/0.356/0.356/0.000 ms 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@448 -- # return 0 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@19 -- # nvmfappstart -m 0x7 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@724 -- # xtrace_disable 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@507 -- # nvmfpid=1702724 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@508 -- # waitforlisten 1702724 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@831 -- # '[' -z 1702724 ']' 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@836 -- # local max_retries=100 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:40.806 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@840 -- # xtrace_disable 00:10:40.806 08:46:30 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:10:40.806 [2024-11-06 08:46:30.781054] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:10:40.806 [2024-11-06 08:46:30.781121] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:40.806 [2024-11-06 08:46:30.865665] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:10:40.806 [2024-11-06 08:46:30.907243] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:40.806 [2024-11-06 08:46:30.907284] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:40.806 [2024-11-06 08:46:30.907295] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:40.806 [2024-11-06 08:46:30.907302] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:40.806 [2024-11-06 08:46:30.907308] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:40.806 [2024-11-06 08:46:30.908732] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:10:40.806 [2024-11-06 08:46:30.908876] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:10:40.806 [2024-11-06 08:46:30.908973] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:10:41.748 08:46:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:10:41.748 08:46:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@864 -- # return 0 00:10:41.748 08:46:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:10:41.748 08:46:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@730 -- # xtrace_disable 00:10:41.748 08:46:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:10:41.748 08:46:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:41.748 08:46:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:10:41.748 [2024-11-06 08:46:31.784869] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:41.748 08:46:31 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:10:42.009 08:46:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # base_bdevs='Malloc0 ' 00:10:42.009 08:46:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:10:42.270 08:46:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # base_bdevs+=Malloc1 00:10:42.270 08:46:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc0 Malloc1' 00:10:42.531 08:46:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore raid0 lvs 00:10:42.531 08:46:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # lvs=eee842ee-46d8-4111-81e4-edfd3409b13b 00:10:42.531 08:46:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u eee842ee-46d8-4111-81e4-edfd3409b13b lvol 20 00:10:42.792 08:46:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # lvol=3c43078d-1cd0-496d-8555-fdaab6ec6ca8 00:10:42.792 08:46:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:10:43.053 08:46:32 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 3c43078d-1cd0-496d-8555-fdaab6ec6ca8 00:10:43.053 08:46:33 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:10:43.315 [2024-11-06 08:46:33.303002] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:43.315 08:46:33 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:10:43.575 08:46:33 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@42 -- # perf_pid=1703433 00:10:43.575 08:46:33 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@44 -- # sleep 1 00:10:43.575 08:46:33 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -o 4096 -q 128 -s 512 -w randwrite -t 10 -c 0x18 00:10:44.515 08:46:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_snapshot 3c43078d-1cd0-496d-8555-fdaab6ec6ca8 MY_SNAPSHOT 00:10:44.776 08:46:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # snapshot=7bae8e52-1319-41a3-918a-4448644ce2c4 00:10:44.776 08:46:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_resize 3c43078d-1cd0-496d-8555-fdaab6ec6ca8 30 00:10:45.039 08:46:34 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_clone 7bae8e52-1319-41a3-918a-4448644ce2c4 MY_CLONE 00:10:45.300 08:46:35 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # clone=93d1d725-62ce-4239-b721-14a4c26915bc 00:10:45.300 08:46:35 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_inflate 93d1d725-62ce-4239-b721-14a4c26915bc 00:10:45.561 08:46:35 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@53 -- # wait 1703433 00:10:53.691 Initializing NVMe Controllers 00:10:53.691 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:10:53.691 Controller IO queue size 128, less than required. 00:10:53.691 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:10:53.691 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 3 00:10:53.691 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 4 00:10:53.691 Initialization complete. Launching workers. 00:10:53.691 ======================================================== 00:10:53.691 Latency(us) 00:10:53.691 Device Information : IOPS MiB/s Average min max 00:10:53.691 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 3: 12176.80 47.57 10514.34 1516.75 56874.67 00:10:53.691 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 4: 17728.79 69.25 7220.47 500.22 50833.29 00:10:53.691 ======================================================== 00:10:53.692 Total : 29905.59 116.82 8561.65 500.22 56874.67 00:10:53.692 00:10:53.952 08:46:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:10:53.952 08:46:43 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 3c43078d-1cd0-496d-8555-fdaab6ec6ca8 00:10:54.211 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u eee842ee-46d8-4111-81e4-edfd3409b13b 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@60 -- # rm -f 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@62 -- # trap - SIGINT SIGTERM EXIT 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- target/nvmf_lvol.sh@64 -- # nvmftestfini 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@514 -- # nvmfcleanup 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@121 -- # sync 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@124 -- # set +e 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@125 -- # for i in {1..20} 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:10:54.471 rmmod nvme_tcp 00:10:54.471 rmmod nvme_fabrics 00:10:54.471 rmmod nvme_keyring 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@128 -- # set -e 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@129 -- # return 0 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@515 -- # '[' -n 1702724 ']' 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@516 -- # killprocess 1702724 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@950 -- # '[' -z 1702724 ']' 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@954 -- # kill -0 1702724 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@955 -- # uname 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1702724 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1702724' 00:10:54.471 killing process with pid 1702724 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@969 -- # kill 1702724 00:10:54.471 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@974 -- # wait 1702724 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@297 -- # iptr 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@789 -- # iptables-save 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@789 -- # iptables-restore 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@302 -- # remove_spdk_ns 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:54.731 08:46:44 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:56.641 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvol -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:10:56.641 00:10:56.641 real 0m23.200s 00:10:56.641 user 1m4.000s 00:10:56.641 sys 0m8.236s 00:10:56.641 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@1126 -- # xtrace_disable 00:10:56.641 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:10:56.641 ************************************ 00:10:56.641 END TEST nvmf_lvol 00:10:56.641 ************************************ 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@28 -- # run_test nvmf_lvs_grow /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:10:56.902 ************************************ 00:10:56.902 START TEST nvmf_lvs_grow 00:10:56.902 ************************************ 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:10:56.902 * Looking for test storage... 00:10:56.902 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1689 -- # lcov --version 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@333 -- # local ver1 ver1_l 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@334 -- # local ver2 ver2_l 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@336 -- # IFS=.-: 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@336 -- # read -ra ver1 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@337 -- # IFS=.-: 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@337 -- # read -ra ver2 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@338 -- # local 'op=<' 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@340 -- # ver1_l=2 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@341 -- # ver2_l=1 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@344 -- # case "$op" in 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@345 -- # : 1 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v = 0 )) 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@365 -- # decimal 1 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=1 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:10:56.902 08:46:46 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 1 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@365 -- # ver1[v]=1 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@366 -- # decimal 2 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=2 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 2 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@366 -- # ver2[v]=2 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@368 -- # return 0 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:10:56.902 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.902 --rc genhtml_branch_coverage=1 00:10:56.902 --rc genhtml_function_coverage=1 00:10:56.902 --rc genhtml_legend=1 00:10:56.902 --rc geninfo_all_blocks=1 00:10:56.902 --rc geninfo_unexecuted_blocks=1 00:10:56.902 00:10:56.902 ' 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:10:56.902 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.902 --rc genhtml_branch_coverage=1 00:10:56.902 --rc genhtml_function_coverage=1 00:10:56.902 --rc genhtml_legend=1 00:10:56.902 --rc geninfo_all_blocks=1 00:10:56.902 --rc geninfo_unexecuted_blocks=1 00:10:56.902 00:10:56.902 ' 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:10:56.902 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.902 --rc genhtml_branch_coverage=1 00:10:56.902 --rc genhtml_function_coverage=1 00:10:56.902 --rc genhtml_legend=1 00:10:56.902 --rc geninfo_all_blocks=1 00:10:56.902 --rc geninfo_unexecuted_blocks=1 00:10:56.902 00:10:56.902 ' 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:10:56.902 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:10:56.902 --rc genhtml_branch_coverage=1 00:10:56.902 --rc genhtml_function_coverage=1 00:10:56.902 --rc genhtml_legend=1 00:10:56.902 --rc geninfo_all_blocks=1 00:10:56.902 --rc geninfo_unexecuted_blocks=1 00:10:56.902 00:10:56.902 ' 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:56.902 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@7 -- # uname -s 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@15 -- # shopt -s extglob 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:57.163 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@5 -- # export PATH 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@51 -- # : 0 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:10:57.164 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@55 -- # have_pci_nics=0 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@12 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@98 -- # nvmftestinit 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@474 -- # prepare_net_devs 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@436 -- # local -g is_hw=no 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@438 -- # remove_spdk_ns 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@309 -- # xtrace_disable 00:10:57.164 08:46:47 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@315 -- # pci_devs=() 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@319 -- # net_devs=() 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@320 -- # e810=() 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@320 -- # local -ga e810 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@321 -- # x722=() 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@321 -- # local -ga x722 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@322 -- # mlx=() 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@322 -- # local -ga mlx 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:11:05.302 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:11:05.302 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:05.302 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ up == up ]] 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:11:05.303 Found net devices under 0000:4b:00.0: cvl_0_0 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ up == up ]] 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:11:05.303 Found net devices under 0000:4b:00.1: cvl_0_1 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@440 -- # is_hw=yes 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:05.303 08:46:53 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:05.303 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:05.303 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.545 ms 00:11:05.303 00:11:05.303 --- 10.0.0.2 ping statistics --- 00:11:05.303 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:05.303 rtt min/avg/max/mdev = 0.545/0.545/0.545/0.000 ms 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:05.303 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:05.303 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.063 ms 00:11:05.303 00:11:05.303 --- 10.0.0.1 ping statistics --- 00:11:05.303 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:05.303 rtt min/avg/max/mdev = 0.063/0.063/0.063/0.000 ms 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@448 -- # return 0 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@99 -- # nvmfappstart -m 0x1 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@724 -- # xtrace_disable 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@507 -- # nvmfpid=1709800 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@508 -- # waitforlisten 1709800 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@831 -- # '[' -z 1709800 ']' 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:05.303 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:05.303 08:46:54 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:11:05.303 [2024-11-06 08:46:54.383020] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:11:05.303 [2024-11-06 08:46:54.383077] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:05.303 [2024-11-06 08:46:54.460906] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:05.303 [2024-11-06 08:46:54.497733] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:05.303 [2024-11-06 08:46:54.497772] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:05.303 [2024-11-06 08:46:54.497780] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:05.303 [2024-11-06 08:46:54.497787] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:05.303 [2024-11-06 08:46:54.497792] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:05.303 [2024-11-06 08:46:54.498349] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@864 -- # return 0 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@730 -- # xtrace_disable 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:11:05.303 [2024-11-06 08:46:55.348190] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@102 -- # run_test lvs_grow_clean lvs_grow 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:05.303 ************************************ 00:11:05.303 START TEST lvs_grow_clean 00:11:05.303 ************************************ 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1125 -- # lvs_grow 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:11:05.303 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:11:05.304 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:11:05.304 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:11:05.304 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:05.304 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:05.304 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:11:05.564 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:11:05.564 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:11:05.825 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # lvs=5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:05.825 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:05.825 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:11:06.085 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:11:06.085 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:11:06.085 08:46:55 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 5004793e-f364-43ce-a0aa-24147fe2c08d lvol 150 00:11:06.085 08:46:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # lvol=ef4f9b06-511f-4e14-bcfc-c73ceeb8a002 00:11:06.085 08:46:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:06.085 08:46:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:11:06.346 [2024-11-06 08:46:56.260224] bdev_aio.c:1044:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:11:06.346 [2024-11-06 08:46:56.260278] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:11:06.346 true 00:11:06.346 08:46:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:06.346 08:46:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:11:06.346 08:46:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:11:06.346 08:46:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:11:06.606 08:46:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 ef4f9b06-511f-4e14-bcfc-c73ceeb8a002 00:11:06.866 08:46:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:11:06.866 [2024-11-06 08:46:56.950326] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:06.866 08:46:56 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:11:07.126 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=1710312 00:11:07.126 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:11:07.126 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 1710312 /var/tmp/bdevperf.sock 00:11:07.126 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@831 -- # '[' -z 1710312 ']' 00:11:07.126 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:11:07.126 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:07.126 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:11:07.126 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:11:07.126 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:07.126 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:11:07.126 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:11:07.126 [2024-11-06 08:46:57.184291] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:11:07.126 [2024-11-06 08:46:57.184346] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1710312 ] 00:11:07.387 [2024-11-06 08:46:57.271403] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:07.387 [2024-11-06 08:46:57.307534] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:07.959 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:07.959 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@864 -- # return 0 00:11:07.959 08:46:57 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:11:08.219 Nvme0n1 00:11:08.480 08:46:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:11:08.480 [ 00:11:08.480 { 00:11:08.480 "name": "Nvme0n1", 00:11:08.480 "aliases": [ 00:11:08.480 "ef4f9b06-511f-4e14-bcfc-c73ceeb8a002" 00:11:08.480 ], 00:11:08.480 "product_name": "NVMe disk", 00:11:08.480 "block_size": 4096, 00:11:08.480 "num_blocks": 38912, 00:11:08.480 "uuid": "ef4f9b06-511f-4e14-bcfc-c73ceeb8a002", 00:11:08.480 "numa_id": 0, 00:11:08.480 "assigned_rate_limits": { 00:11:08.480 "rw_ios_per_sec": 0, 00:11:08.480 "rw_mbytes_per_sec": 0, 00:11:08.480 "r_mbytes_per_sec": 0, 00:11:08.480 "w_mbytes_per_sec": 0 00:11:08.480 }, 00:11:08.480 "claimed": false, 00:11:08.480 "zoned": false, 00:11:08.480 "supported_io_types": { 00:11:08.480 "read": true, 00:11:08.480 "write": true, 00:11:08.480 "unmap": true, 00:11:08.480 "flush": true, 00:11:08.480 "reset": true, 00:11:08.480 "nvme_admin": true, 00:11:08.480 "nvme_io": true, 00:11:08.480 "nvme_io_md": false, 00:11:08.480 "write_zeroes": true, 00:11:08.480 "zcopy": false, 00:11:08.480 "get_zone_info": false, 00:11:08.480 "zone_management": false, 00:11:08.480 "zone_append": false, 00:11:08.480 "compare": true, 00:11:08.480 "compare_and_write": true, 00:11:08.480 "abort": true, 00:11:08.480 "seek_hole": false, 00:11:08.480 "seek_data": false, 00:11:08.480 "copy": true, 00:11:08.480 "nvme_iov_md": false 00:11:08.480 }, 00:11:08.480 "memory_domains": [ 00:11:08.480 { 00:11:08.480 "dma_device_id": "system", 00:11:08.480 "dma_device_type": 1 00:11:08.480 } 00:11:08.480 ], 00:11:08.480 "driver_specific": { 00:11:08.480 "nvme": [ 00:11:08.480 { 00:11:08.480 "trid": { 00:11:08.480 "trtype": "TCP", 00:11:08.480 "adrfam": "IPv4", 00:11:08.480 "traddr": "10.0.0.2", 00:11:08.480 "trsvcid": "4420", 00:11:08.480 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:11:08.480 }, 00:11:08.480 "ctrlr_data": { 00:11:08.480 "cntlid": 1, 00:11:08.480 "vendor_id": "0x8086", 00:11:08.480 "model_number": "SPDK bdev Controller", 00:11:08.480 "serial_number": "SPDK0", 00:11:08.480 "firmware_revision": "25.01", 00:11:08.480 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:11:08.480 "oacs": { 00:11:08.480 "security": 0, 00:11:08.480 "format": 0, 00:11:08.480 "firmware": 0, 00:11:08.480 "ns_manage": 0 00:11:08.480 }, 00:11:08.480 "multi_ctrlr": true, 00:11:08.480 "ana_reporting": false 00:11:08.480 }, 00:11:08.480 "vs": { 00:11:08.480 "nvme_version": "1.3" 00:11:08.480 }, 00:11:08.480 "ns_data": { 00:11:08.480 "id": 1, 00:11:08.480 "can_share": true 00:11:08.480 } 00:11:08.480 } 00:11:08.480 ], 00:11:08.480 "mp_policy": "active_passive" 00:11:08.480 } 00:11:08.480 } 00:11:08.480 ] 00:11:08.480 08:46:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=1710537 00:11:08.480 08:46:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:11:08.480 08:46:58 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:11:08.480 Running I/O for 10 seconds... 00:11:09.862 Latency(us) 00:11:09.862 [2024-11-06T07:46:59.975Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:09.862 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:09.862 Nvme0n1 : 1.00 17718.00 69.21 0.00 0.00 0.00 0.00 0.00 00:11:09.862 [2024-11-06T07:46:59.975Z] =================================================================================================================== 00:11:09.862 [2024-11-06T07:46:59.975Z] Total : 17718.00 69.21 0.00 0.00 0.00 0.00 0.00 00:11:09.862 00:11:10.433 08:47:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:10.693 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:10.693 Nvme0n1 : 2.00 17807.50 69.56 0.00 0.00 0.00 0.00 0.00 00:11:10.693 [2024-11-06T07:47:00.806Z] =================================================================================================================== 00:11:10.693 [2024-11-06T07:47:00.806Z] Total : 17807.50 69.56 0.00 0.00 0.00 0.00 0.00 00:11:10.693 00:11:10.693 true 00:11:10.694 08:47:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:10.694 08:47:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:11:10.953 08:47:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:11:10.953 08:47:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:11:10.953 08:47:00 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@65 -- # wait 1710537 00:11:11.523 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:11.523 Nvme0n1 : 3.00 17852.00 69.73 0.00 0.00 0.00 0.00 0.00 00:11:11.523 [2024-11-06T07:47:01.636Z] =================================================================================================================== 00:11:11.523 [2024-11-06T07:47:01.636Z] Total : 17852.00 69.73 0.00 0.00 0.00 0.00 0.00 00:11:11.523 00:11:12.908 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:12.908 Nvme0n1 : 4.00 17880.50 69.85 0.00 0.00 0.00 0.00 0.00 00:11:12.908 [2024-11-06T07:47:03.021Z] =================================================================================================================== 00:11:12.908 [2024-11-06T07:47:03.021Z] Total : 17880.50 69.85 0.00 0.00 0.00 0.00 0.00 00:11:12.908 00:11:13.850 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:13.850 Nvme0n1 : 5.00 17922.00 70.01 0.00 0.00 0.00 0.00 0.00 00:11:13.850 [2024-11-06T07:47:03.963Z] =================================================================================================================== 00:11:13.850 [2024-11-06T07:47:03.963Z] Total : 17922.00 70.01 0.00 0.00 0.00 0.00 0.00 00:11:13.850 00:11:14.792 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:14.792 Nvme0n1 : 6.00 17938.83 70.07 0.00 0.00 0.00 0.00 0.00 00:11:14.792 [2024-11-06T07:47:04.905Z] =================================================================================================================== 00:11:14.792 [2024-11-06T07:47:04.905Z] Total : 17938.83 70.07 0.00 0.00 0.00 0.00 0.00 00:11:14.792 00:11:15.733 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:15.733 Nvme0n1 : 7.00 17960.86 70.16 0.00 0.00 0.00 0.00 0.00 00:11:15.733 [2024-11-06T07:47:05.846Z] =================================================================================================================== 00:11:15.733 [2024-11-06T07:47:05.846Z] Total : 17960.86 70.16 0.00 0.00 0.00 0.00 0.00 00:11:15.733 00:11:16.740 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:16.740 Nvme0n1 : 8.00 17972.00 70.20 0.00 0.00 0.00 0.00 0.00 00:11:16.740 [2024-11-06T07:47:06.853Z] =================================================================================================================== 00:11:16.740 [2024-11-06T07:47:06.853Z] Total : 17972.00 70.20 0.00 0.00 0.00 0.00 0.00 00:11:16.740 00:11:17.680 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:17.680 Nvme0n1 : 9.00 17984.44 70.25 0.00 0.00 0.00 0.00 0.00 00:11:17.680 [2024-11-06T07:47:07.793Z] =================================================================================================================== 00:11:17.680 [2024-11-06T07:47:07.793Z] Total : 17984.44 70.25 0.00 0.00 0.00 0.00 0.00 00:11:17.680 00:11:18.626 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:18.626 Nvme0n1 : 10.00 17994.40 70.29 0.00 0.00 0.00 0.00 0.00 00:11:18.626 [2024-11-06T07:47:08.739Z] =================================================================================================================== 00:11:18.626 [2024-11-06T07:47:08.739Z] Total : 17994.40 70.29 0.00 0.00 0.00 0.00 0.00 00:11:18.626 00:11:18.626 00:11:18.626 Latency(us) 00:11:18.626 [2024-11-06T07:47:08.739Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:18.626 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:18.626 Nvme0n1 : 10.00 17998.27 70.31 0.00 0.00 7107.65 2225.49 13544.11 00:11:18.626 [2024-11-06T07:47:08.739Z] =================================================================================================================== 00:11:18.626 [2024-11-06T07:47:08.739Z] Total : 17998.27 70.31 0.00 0.00 7107.65 2225.49 13544.11 00:11:18.626 { 00:11:18.626 "results": [ 00:11:18.626 { 00:11:18.626 "job": "Nvme0n1", 00:11:18.626 "core_mask": "0x2", 00:11:18.626 "workload": "randwrite", 00:11:18.626 "status": "finished", 00:11:18.626 "queue_depth": 128, 00:11:18.626 "io_size": 4096, 00:11:18.626 "runtime": 10.004963, 00:11:18.626 "iops": 17998.26745985967, 00:11:18.626 "mibps": 70.30573226507684, 00:11:18.626 "io_failed": 0, 00:11:18.626 "io_timeout": 0, 00:11:18.626 "avg_latency_us": 7107.647242732537, 00:11:18.626 "min_latency_us": 2225.4933333333333, 00:11:18.626 "max_latency_us": 13544.106666666667 00:11:18.626 } 00:11:18.626 ], 00:11:18.626 "core_count": 1 00:11:18.626 } 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@66 -- # killprocess 1710312 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@950 -- # '[' -z 1710312 ']' 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@954 -- # kill -0 1710312 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@955 -- # uname 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1710312 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1710312' 00:11:18.626 killing process with pid 1710312 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@969 -- # kill 1710312 00:11:18.626 Received shutdown signal, test time was about 10.000000 seconds 00:11:18.626 00:11:18.626 Latency(us) 00:11:18.626 [2024-11-06T07:47:08.739Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:18.626 [2024-11-06T07:47:08.739Z] =================================================================================================================== 00:11:18.626 [2024-11-06T07:47:08.739Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:11:18.626 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@974 -- # wait 1710312 00:11:18.887 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:11:18.887 08:47:08 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:11:19.147 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:19.147 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:11:19.407 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:11:19.407 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@72 -- # [[ '' == \d\i\r\t\y ]] 00:11:19.407 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:11:19.407 [2024-11-06 08:47:09.445929] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:11:19.407 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@650 -- # local es=0 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:11:19.408 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:19.667 request: 00:11:19.667 { 00:11:19.667 "uuid": "5004793e-f364-43ce-a0aa-24147fe2c08d", 00:11:19.668 "method": "bdev_lvol_get_lvstores", 00:11:19.668 "req_id": 1 00:11:19.668 } 00:11:19.668 Got JSON-RPC error response 00:11:19.668 response: 00:11:19.668 { 00:11:19.668 "code": -19, 00:11:19.668 "message": "No such device" 00:11:19.668 } 00:11:19.668 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@653 -- # es=1 00:11:19.668 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:11:19.668 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:11:19.668 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:11:19.668 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:11:19.927 aio_bdev 00:11:19.927 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev ef4f9b06-511f-4e14-bcfc-c73ceeb8a002 00:11:19.927 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@899 -- # local bdev_name=ef4f9b06-511f-4e14-bcfc-c73ceeb8a002 00:11:19.927 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:19.927 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@901 -- # local i 00:11:19.927 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:19.927 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:19.927 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:11:19.927 08:47:09 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b ef4f9b06-511f-4e14-bcfc-c73ceeb8a002 -t 2000 00:11:20.187 [ 00:11:20.187 { 00:11:20.187 "name": "ef4f9b06-511f-4e14-bcfc-c73ceeb8a002", 00:11:20.187 "aliases": [ 00:11:20.187 "lvs/lvol" 00:11:20.187 ], 00:11:20.187 "product_name": "Logical Volume", 00:11:20.187 "block_size": 4096, 00:11:20.187 "num_blocks": 38912, 00:11:20.187 "uuid": "ef4f9b06-511f-4e14-bcfc-c73ceeb8a002", 00:11:20.187 "assigned_rate_limits": { 00:11:20.187 "rw_ios_per_sec": 0, 00:11:20.187 "rw_mbytes_per_sec": 0, 00:11:20.187 "r_mbytes_per_sec": 0, 00:11:20.187 "w_mbytes_per_sec": 0 00:11:20.187 }, 00:11:20.187 "claimed": false, 00:11:20.187 "zoned": false, 00:11:20.187 "supported_io_types": { 00:11:20.187 "read": true, 00:11:20.187 "write": true, 00:11:20.187 "unmap": true, 00:11:20.187 "flush": false, 00:11:20.187 "reset": true, 00:11:20.187 "nvme_admin": false, 00:11:20.187 "nvme_io": false, 00:11:20.187 "nvme_io_md": false, 00:11:20.187 "write_zeroes": true, 00:11:20.187 "zcopy": false, 00:11:20.187 "get_zone_info": false, 00:11:20.187 "zone_management": false, 00:11:20.187 "zone_append": false, 00:11:20.187 "compare": false, 00:11:20.187 "compare_and_write": false, 00:11:20.187 "abort": false, 00:11:20.187 "seek_hole": true, 00:11:20.187 "seek_data": true, 00:11:20.187 "copy": false, 00:11:20.187 "nvme_iov_md": false 00:11:20.187 }, 00:11:20.187 "driver_specific": { 00:11:20.187 "lvol": { 00:11:20.187 "lvol_store_uuid": "5004793e-f364-43ce-a0aa-24147fe2c08d", 00:11:20.187 "base_bdev": "aio_bdev", 00:11:20.187 "thin_provision": false, 00:11:20.187 "num_allocated_clusters": 38, 00:11:20.187 "snapshot": false, 00:11:20.187 "clone": false, 00:11:20.187 "esnap_clone": false 00:11:20.187 } 00:11:20.187 } 00:11:20.187 } 00:11:20.187 ] 00:11:20.187 08:47:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@907 -- # return 0 00:11:20.187 08:47:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:20.187 08:47:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:11:20.447 08:47:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:11:20.447 08:47:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:20.448 08:47:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:11:20.448 08:47:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:11:20.448 08:47:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete ef4f9b06-511f-4e14-bcfc-c73ceeb8a002 00:11:20.708 08:47:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 5004793e-f364-43ce-a0aa-24147fe2c08d 00:11:20.969 08:47:10 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:11:20.969 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:20.969 00:11:20.969 real 0m15.665s 00:11:20.969 user 0m15.343s 00:11:20.969 sys 0m1.372s 00:11:20.969 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:20.969 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:11:20.969 ************************************ 00:11:20.969 END TEST lvs_grow_clean 00:11:20.969 ************************************ 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@103 -- # run_test lvs_grow_dirty lvs_grow dirty 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:21.230 ************************************ 00:11:21.230 START TEST lvs_grow_dirty 00:11:21.230 ************************************ 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1125 -- # lvs_grow dirty 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:11:21.230 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:11:21.490 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # lvs=89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:21.490 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:21.490 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:11:21.751 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:11:21.751 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:11:21.751 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 lvol 150 00:11:21.751 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # lvol=43ae1d17-e8a8-4d14-9f51-13633cf94601 00:11:21.751 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:21.751 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:11:22.012 [2024-11-06 08:47:11.975429] bdev_aio.c:1044:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:11:22.012 [2024-11-06 08:47:11.975480] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:11:22.012 true 00:11:22.012 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:22.012 08:47:11 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:11:22.278 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:11:22.278 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:11:22.278 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 43ae1d17-e8a8-4d14-9f51-13633cf94601 00:11:22.539 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:11:22.539 [2024-11-06 08:47:12.629404] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:22.539 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:11:22.799 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=1713610 00:11:22.799 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:11:22.799 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:11:22.799 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 1713610 /var/tmp/bdevperf.sock 00:11:22.799 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@831 -- # '[' -z 1713610 ']' 00:11:22.799 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:11:22.799 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:22.799 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:11:22.799 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:11:22.799 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:22.799 08:47:12 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:11:22.799 [2024-11-06 08:47:12.862842] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:11:22.799 [2024-11-06 08:47:12.862898] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1713610 ] 00:11:23.059 [2024-11-06 08:47:12.949874] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:23.059 [2024-11-06 08:47:12.985937] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:23.629 08:47:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:23.629 08:47:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # return 0 00:11:23.629 08:47:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:11:23.889 Nvme0n1 00:11:23.889 08:47:13 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:11:24.150 [ 00:11:24.150 { 00:11:24.150 "name": "Nvme0n1", 00:11:24.150 "aliases": [ 00:11:24.150 "43ae1d17-e8a8-4d14-9f51-13633cf94601" 00:11:24.150 ], 00:11:24.150 "product_name": "NVMe disk", 00:11:24.150 "block_size": 4096, 00:11:24.150 "num_blocks": 38912, 00:11:24.150 "uuid": "43ae1d17-e8a8-4d14-9f51-13633cf94601", 00:11:24.150 "numa_id": 0, 00:11:24.150 "assigned_rate_limits": { 00:11:24.150 "rw_ios_per_sec": 0, 00:11:24.150 "rw_mbytes_per_sec": 0, 00:11:24.150 "r_mbytes_per_sec": 0, 00:11:24.150 "w_mbytes_per_sec": 0 00:11:24.150 }, 00:11:24.150 "claimed": false, 00:11:24.150 "zoned": false, 00:11:24.150 "supported_io_types": { 00:11:24.150 "read": true, 00:11:24.150 "write": true, 00:11:24.150 "unmap": true, 00:11:24.150 "flush": true, 00:11:24.150 "reset": true, 00:11:24.150 "nvme_admin": true, 00:11:24.150 "nvme_io": true, 00:11:24.150 "nvme_io_md": false, 00:11:24.150 "write_zeroes": true, 00:11:24.150 "zcopy": false, 00:11:24.150 "get_zone_info": false, 00:11:24.150 "zone_management": false, 00:11:24.150 "zone_append": false, 00:11:24.150 "compare": true, 00:11:24.150 "compare_and_write": true, 00:11:24.150 "abort": true, 00:11:24.150 "seek_hole": false, 00:11:24.150 "seek_data": false, 00:11:24.150 "copy": true, 00:11:24.150 "nvme_iov_md": false 00:11:24.150 }, 00:11:24.150 "memory_domains": [ 00:11:24.150 { 00:11:24.150 "dma_device_id": "system", 00:11:24.150 "dma_device_type": 1 00:11:24.150 } 00:11:24.150 ], 00:11:24.150 "driver_specific": { 00:11:24.150 "nvme": [ 00:11:24.150 { 00:11:24.150 "trid": { 00:11:24.150 "trtype": "TCP", 00:11:24.150 "adrfam": "IPv4", 00:11:24.150 "traddr": "10.0.0.2", 00:11:24.150 "trsvcid": "4420", 00:11:24.150 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:11:24.150 }, 00:11:24.150 "ctrlr_data": { 00:11:24.150 "cntlid": 1, 00:11:24.150 "vendor_id": "0x8086", 00:11:24.150 "model_number": "SPDK bdev Controller", 00:11:24.150 "serial_number": "SPDK0", 00:11:24.150 "firmware_revision": "25.01", 00:11:24.150 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:11:24.150 "oacs": { 00:11:24.150 "security": 0, 00:11:24.150 "format": 0, 00:11:24.150 "firmware": 0, 00:11:24.150 "ns_manage": 0 00:11:24.150 }, 00:11:24.150 "multi_ctrlr": true, 00:11:24.150 "ana_reporting": false 00:11:24.150 }, 00:11:24.150 "vs": { 00:11:24.150 "nvme_version": "1.3" 00:11:24.150 }, 00:11:24.150 "ns_data": { 00:11:24.150 "id": 1, 00:11:24.150 "can_share": true 00:11:24.150 } 00:11:24.150 } 00:11:24.150 ], 00:11:24.150 "mp_policy": "active_passive" 00:11:24.150 } 00:11:24.150 } 00:11:24.150 ] 00:11:24.150 08:47:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=1713748 00:11:24.150 08:47:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:11:24.150 08:47:14 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:11:24.150 Running I/O for 10 seconds... 00:11:25.090 Latency(us) 00:11:25.090 [2024-11-06T07:47:15.203Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:25.090 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:25.090 Nvme0n1 : 1.00 17643.00 68.92 0.00 0.00 0.00 0.00 0.00 00:11:25.090 [2024-11-06T07:47:15.203Z] =================================================================================================================== 00:11:25.090 [2024-11-06T07:47:15.203Z] Total : 17643.00 68.92 0.00 0.00 0.00 0.00 0.00 00:11:25.090 00:11:26.029 08:47:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:26.290 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:26.290 Nvme0n1 : 2.00 17728.00 69.25 0.00 0.00 0.00 0.00 0.00 00:11:26.290 [2024-11-06T07:47:16.403Z] =================================================================================================================== 00:11:26.290 [2024-11-06T07:47:16.403Z] Total : 17728.00 69.25 0.00 0.00 0.00 0.00 0.00 00:11:26.290 00:11:26.290 true 00:11:26.290 08:47:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:26.290 08:47:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:11:26.551 08:47:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:11:26.551 08:47:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:11:26.551 08:47:16 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@65 -- # wait 1713748 00:11:27.121 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:27.121 Nvme0n1 : 3.00 17792.33 69.50 0.00 0.00 0.00 0.00 0.00 00:11:27.121 [2024-11-06T07:47:17.234Z] =================================================================================================================== 00:11:27.121 [2024-11-06T07:47:17.234Z] Total : 17792.33 69.50 0.00 0.00 0.00 0.00 0.00 00:11:27.121 00:11:28.508 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:28.508 Nvme0n1 : 4.00 17848.00 69.72 0.00 0.00 0.00 0.00 0.00 00:11:28.508 [2024-11-06T07:47:18.621Z] =================================================================================================================== 00:11:28.508 [2024-11-06T07:47:18.621Z] Total : 17848.00 69.72 0.00 0.00 0.00 0.00 0.00 00:11:28.508 00:11:29.079 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:29.079 Nvme0n1 : 5.00 17878.60 69.84 0.00 0.00 0.00 0.00 0.00 00:11:29.079 [2024-11-06T07:47:19.192Z] =================================================================================================================== 00:11:29.079 [2024-11-06T07:47:19.192Z] Total : 17878.60 69.84 0.00 0.00 0.00 0.00 0.00 00:11:29.079 00:11:30.468 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:30.468 Nvme0n1 : 6.00 17898.33 69.92 0.00 0.00 0.00 0.00 0.00 00:11:30.468 [2024-11-06T07:47:20.581Z] =================================================================================================================== 00:11:30.468 [2024-11-06T07:47:20.581Z] Total : 17898.33 69.92 0.00 0.00 0.00 0.00 0.00 00:11:30.468 00:11:31.410 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:31.410 Nvme0n1 : 7.00 17920.57 70.00 0.00 0.00 0.00 0.00 0.00 00:11:31.410 [2024-11-06T07:47:21.523Z] =================================================================================================================== 00:11:31.410 [2024-11-06T07:47:21.523Z] Total : 17920.57 70.00 0.00 0.00 0.00 0.00 0.00 00:11:31.410 00:11:32.354 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:32.354 Nvme0n1 : 8.00 17940.62 70.08 0.00 0.00 0.00 0.00 0.00 00:11:32.354 [2024-11-06T07:47:22.467Z] =================================================================================================================== 00:11:32.354 [2024-11-06T07:47:22.467Z] Total : 17940.62 70.08 0.00 0.00 0.00 0.00 0.00 00:11:32.354 00:11:33.296 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:33.296 Nvme0n1 : 9.00 17943.33 70.09 0.00 0.00 0.00 0.00 0.00 00:11:33.297 [2024-11-06T07:47:23.410Z] =================================================================================================================== 00:11:33.297 [2024-11-06T07:47:23.410Z] Total : 17943.33 70.09 0.00 0.00 0.00 0.00 0.00 00:11:33.297 00:11:34.239 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:34.239 Nvme0n1 : 10.00 17964.20 70.17 0.00 0.00 0.00 0.00 0.00 00:11:34.239 [2024-11-06T07:47:24.352Z] =================================================================================================================== 00:11:34.239 [2024-11-06T07:47:24.352Z] Total : 17964.20 70.17 0.00 0.00 0.00 0.00 0.00 00:11:34.239 00:11:34.239 00:11:34.239 Latency(us) 00:11:34.239 [2024-11-06T07:47:24.352Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:34.239 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:11:34.239 Nvme0n1 : 10.01 17964.68 70.17 0.00 0.00 7121.49 4341.76 14090.24 00:11:34.239 [2024-11-06T07:47:24.352Z] =================================================================================================================== 00:11:34.239 [2024-11-06T07:47:24.352Z] Total : 17964.68 70.17 0.00 0.00 7121.49 4341.76 14090.24 00:11:34.239 { 00:11:34.239 "results": [ 00:11:34.239 { 00:11:34.239 "job": "Nvme0n1", 00:11:34.239 "core_mask": "0x2", 00:11:34.239 "workload": "randwrite", 00:11:34.239 "status": "finished", 00:11:34.239 "queue_depth": 128, 00:11:34.239 "io_size": 4096, 00:11:34.239 "runtime": 10.006857, 00:11:34.239 "iops": 17964.681617814666, 00:11:34.239 "mibps": 70.17453756958854, 00:11:34.239 "io_failed": 0, 00:11:34.239 "io_timeout": 0, 00:11:34.239 "avg_latency_us": 7121.489882405296, 00:11:34.239 "min_latency_us": 4341.76, 00:11:34.239 "max_latency_us": 14090.24 00:11:34.239 } 00:11:34.239 ], 00:11:34.239 "core_count": 1 00:11:34.239 } 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@66 -- # killprocess 1713610 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@950 -- # '[' -z 1713610 ']' 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@954 -- # kill -0 1713610 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@955 -- # uname 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1713610 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1713610' 00:11:34.239 killing process with pid 1713610 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@969 -- # kill 1713610 00:11:34.239 Received shutdown signal, test time was about 10.000000 seconds 00:11:34.239 00:11:34.239 Latency(us) 00:11:34.239 [2024-11-06T07:47:24.352Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:34.239 [2024-11-06T07:47:24.352Z] =================================================================================================================== 00:11:34.239 [2024-11-06T07:47:24.352Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:11:34.239 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@974 -- # wait 1713610 00:11:34.500 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:11:34.500 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:11:34.761 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:34.761 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:11:35.022 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:11:35.022 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@72 -- # [[ dirty == \d\i\r\t\y ]] 00:11:35.022 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@74 -- # kill -9 1709800 00:11:35.022 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # wait 1709800 00:11:35.023 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh: line 75: 1709800 Killed "${NVMF_APP[@]}" "$@" 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # true 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@76 -- # nvmfappstart -m 0x1 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@724 -- # xtrace_disable 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@507 -- # nvmfpid=1715986 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@508 -- # waitforlisten 1715986 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@831 -- # '[' -z 1715986 ']' 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:35.023 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:35.023 08:47:24 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:11:35.023 [2024-11-06 08:47:25.008666] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:11:35.023 [2024-11-06 08:47:25.008724] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:35.023 [2024-11-06 08:47:25.085104] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:35.023 [2024-11-06 08:47:25.119641] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:35.023 [2024-11-06 08:47:25.119674] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:35.023 [2024-11-06 08:47:25.119682] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:35.023 [2024-11-06 08:47:25.119688] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:35.023 [2024-11-06 08:47:25.119694] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:35.023 [2024-11-06 08:47:25.120287] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:35.963 08:47:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:35.963 08:47:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # return 0 00:11:35.963 08:47:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:11:35.963 08:47:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@730 -- # xtrace_disable 00:11:35.963 08:47:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:11:35.963 08:47:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:35.963 08:47:25 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:11:35.963 [2024-11-06 08:47:26.000006] blobstore.c:4875:bs_recover: *NOTICE*: Performing recovery on blobstore 00:11:35.963 [2024-11-06 08:47:26.000104] blobstore.c:4822:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x0 00:11:35.963 [2024-11-06 08:47:26.000135] blobstore.c:4822:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x1 00:11:35.963 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # aio_bdev=aio_bdev 00:11:35.963 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@78 -- # waitforbdev 43ae1d17-e8a8-4d14-9f51-13633cf94601 00:11:35.963 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@899 -- # local bdev_name=43ae1d17-e8a8-4d14-9f51-13633cf94601 00:11:35.963 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:35.963 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@901 -- # local i 00:11:35.963 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:35.963 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:35.964 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:11:36.225 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 43ae1d17-e8a8-4d14-9f51-13633cf94601 -t 2000 00:11:36.486 [ 00:11:36.486 { 00:11:36.486 "name": "43ae1d17-e8a8-4d14-9f51-13633cf94601", 00:11:36.486 "aliases": [ 00:11:36.486 "lvs/lvol" 00:11:36.486 ], 00:11:36.486 "product_name": "Logical Volume", 00:11:36.486 "block_size": 4096, 00:11:36.486 "num_blocks": 38912, 00:11:36.486 "uuid": "43ae1d17-e8a8-4d14-9f51-13633cf94601", 00:11:36.486 "assigned_rate_limits": { 00:11:36.486 "rw_ios_per_sec": 0, 00:11:36.486 "rw_mbytes_per_sec": 0, 00:11:36.486 "r_mbytes_per_sec": 0, 00:11:36.486 "w_mbytes_per_sec": 0 00:11:36.486 }, 00:11:36.486 "claimed": false, 00:11:36.486 "zoned": false, 00:11:36.486 "supported_io_types": { 00:11:36.486 "read": true, 00:11:36.486 "write": true, 00:11:36.486 "unmap": true, 00:11:36.486 "flush": false, 00:11:36.486 "reset": true, 00:11:36.486 "nvme_admin": false, 00:11:36.486 "nvme_io": false, 00:11:36.486 "nvme_io_md": false, 00:11:36.486 "write_zeroes": true, 00:11:36.486 "zcopy": false, 00:11:36.486 "get_zone_info": false, 00:11:36.486 "zone_management": false, 00:11:36.486 "zone_append": false, 00:11:36.486 "compare": false, 00:11:36.486 "compare_and_write": false, 00:11:36.486 "abort": false, 00:11:36.486 "seek_hole": true, 00:11:36.486 "seek_data": true, 00:11:36.486 "copy": false, 00:11:36.486 "nvme_iov_md": false 00:11:36.486 }, 00:11:36.486 "driver_specific": { 00:11:36.486 "lvol": { 00:11:36.486 "lvol_store_uuid": "89b23cfc-8b47-45a2-bb6e-e0111ee179e7", 00:11:36.486 "base_bdev": "aio_bdev", 00:11:36.486 "thin_provision": false, 00:11:36.486 "num_allocated_clusters": 38, 00:11:36.486 "snapshot": false, 00:11:36.486 "clone": false, 00:11:36.486 "esnap_clone": false 00:11:36.486 } 00:11:36.486 } 00:11:36.486 } 00:11:36.486 ] 00:11:36.486 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@907 -- # return 0 00:11:36.486 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:36.486 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # jq -r '.[0].free_clusters' 00:11:36.486 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # (( free_clusters == 61 )) 00:11:36.486 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:36.486 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # jq -r '.[0].total_data_clusters' 00:11:36.747 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # (( data_clusters == 99 )) 00:11:36.747 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:11:36.747 [2024-11-06 08:47:26.840154] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@650 -- # local es=0 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:11:37.009 08:47:26 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:37.009 request: 00:11:37.009 { 00:11:37.009 "uuid": "89b23cfc-8b47-45a2-bb6e-e0111ee179e7", 00:11:37.009 "method": "bdev_lvol_get_lvstores", 00:11:37.009 "req_id": 1 00:11:37.009 } 00:11:37.009 Got JSON-RPC error response 00:11:37.009 response: 00:11:37.009 { 00:11:37.009 "code": -19, 00:11:37.009 "message": "No such device" 00:11:37.009 } 00:11:37.009 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@653 -- # es=1 00:11:37.009 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:11:37.009 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:11:37.009 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:11:37.009 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:11:37.270 aio_bdev 00:11:37.270 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev 43ae1d17-e8a8-4d14-9f51-13633cf94601 00:11:37.270 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@899 -- # local bdev_name=43ae1d17-e8a8-4d14-9f51-13633cf94601 00:11:37.270 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:11:37.270 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@901 -- # local i 00:11:37.270 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:11:37.270 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:11:37.270 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:11:37.531 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 43ae1d17-e8a8-4d14-9f51-13633cf94601 -t 2000 00:11:37.531 [ 00:11:37.531 { 00:11:37.531 "name": "43ae1d17-e8a8-4d14-9f51-13633cf94601", 00:11:37.531 "aliases": [ 00:11:37.531 "lvs/lvol" 00:11:37.531 ], 00:11:37.531 "product_name": "Logical Volume", 00:11:37.531 "block_size": 4096, 00:11:37.531 "num_blocks": 38912, 00:11:37.531 "uuid": "43ae1d17-e8a8-4d14-9f51-13633cf94601", 00:11:37.531 "assigned_rate_limits": { 00:11:37.531 "rw_ios_per_sec": 0, 00:11:37.531 "rw_mbytes_per_sec": 0, 00:11:37.531 "r_mbytes_per_sec": 0, 00:11:37.531 "w_mbytes_per_sec": 0 00:11:37.531 }, 00:11:37.531 "claimed": false, 00:11:37.531 "zoned": false, 00:11:37.531 "supported_io_types": { 00:11:37.531 "read": true, 00:11:37.531 "write": true, 00:11:37.531 "unmap": true, 00:11:37.531 "flush": false, 00:11:37.531 "reset": true, 00:11:37.531 "nvme_admin": false, 00:11:37.531 "nvme_io": false, 00:11:37.531 "nvme_io_md": false, 00:11:37.531 "write_zeroes": true, 00:11:37.531 "zcopy": false, 00:11:37.531 "get_zone_info": false, 00:11:37.531 "zone_management": false, 00:11:37.531 "zone_append": false, 00:11:37.531 "compare": false, 00:11:37.531 "compare_and_write": false, 00:11:37.531 "abort": false, 00:11:37.531 "seek_hole": true, 00:11:37.531 "seek_data": true, 00:11:37.531 "copy": false, 00:11:37.531 "nvme_iov_md": false 00:11:37.531 }, 00:11:37.531 "driver_specific": { 00:11:37.531 "lvol": { 00:11:37.531 "lvol_store_uuid": "89b23cfc-8b47-45a2-bb6e-e0111ee179e7", 00:11:37.531 "base_bdev": "aio_bdev", 00:11:37.531 "thin_provision": false, 00:11:37.531 "num_allocated_clusters": 38, 00:11:37.531 "snapshot": false, 00:11:37.531 "clone": false, 00:11:37.531 "esnap_clone": false 00:11:37.531 } 00:11:37.531 } 00:11:37.531 } 00:11:37.531 ] 00:11:37.531 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@907 -- # return 0 00:11:37.531 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:11:37.531 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:37.792 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:11:37.792 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:37.792 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:11:38.053 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:11:38.053 08:47:27 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 43ae1d17-e8a8-4d14-9f51-13633cf94601 00:11:38.053 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 89b23cfc-8b47-45a2-bb6e-e0111ee179e7 00:11:38.313 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:11:38.574 00:11:38.574 real 0m17.383s 00:11:38.574 user 0m45.348s 00:11:38.574 sys 0m2.912s 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:11:38.574 ************************************ 00:11:38.574 END TEST lvs_grow_dirty 00:11:38.574 ************************************ 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # process_shm --id 0 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@808 -- # type=--id 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@809 -- # id=0 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@810 -- # '[' --id = --pid ']' 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # shm_files=nvmf_trace.0 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@816 -- # [[ -z nvmf_trace.0 ]] 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@820 -- # for n in $shm_files 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@821 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:11:38.574 nvmf_trace.0 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@823 -- # return 0 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # nvmftestfini 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@514 -- # nvmfcleanup 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@121 -- # sync 00:11:38.574 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@124 -- # set +e 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:38.575 rmmod nvme_tcp 00:11:38.575 rmmod nvme_fabrics 00:11:38.575 rmmod nvme_keyring 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@128 -- # set -e 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@129 -- # return 0 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@515 -- # '[' -n 1715986 ']' 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@516 -- # killprocess 1715986 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@950 -- # '[' -z 1715986 ']' 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@954 -- # kill -0 1715986 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@955 -- # uname 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:38.575 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1715986 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1715986' 00:11:38.835 killing process with pid 1715986 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@969 -- # kill 1715986 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@974 -- # wait 1715986 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@297 -- # iptr 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@789 -- # iptables-save 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@789 -- # iptables-restore 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:38.835 08:47:28 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:41.377 08:47:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:11:41.377 00:11:41.377 real 0m44.135s 00:11:41.377 user 1m7.015s 00:11:41.377 sys 0m10.211s 00:11:41.377 08:47:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:41.377 08:47:30 nvmf_tcp.nvmf_target_core.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:11:41.377 ************************************ 00:11:41.377 END TEST nvmf_lvs_grow 00:11:41.377 ************************************ 00:11:41.377 08:47:30 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@29 -- # run_test nvmf_bdev_io_wait /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:11:41.377 08:47:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:11:41.377 08:47:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:41.377 08:47:30 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:11:41.377 ************************************ 00:11:41.377 START TEST nvmf_bdev_io_wait 00:11:41.377 ************************************ 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:11:41.377 * Looking for test storage... 00:11:41.377 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1689 -- # lcov --version 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@333 -- # local ver1 ver1_l 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@334 -- # local ver2 ver2_l 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # IFS=.-: 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # read -ra ver1 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # IFS=.-: 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # read -ra ver2 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@338 -- # local 'op=<' 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@340 -- # ver1_l=2 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@341 -- # ver2_l=1 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@344 -- # case "$op" in 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@345 -- # : 1 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v = 0 )) 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # decimal 1 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=1 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 1 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # ver1[v]=1 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # decimal 2 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=2 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 2 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # ver2[v]=2 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # return 0 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:11:41.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:41.377 --rc genhtml_branch_coverage=1 00:11:41.377 --rc genhtml_function_coverage=1 00:11:41.377 --rc genhtml_legend=1 00:11:41.377 --rc geninfo_all_blocks=1 00:11:41.377 --rc geninfo_unexecuted_blocks=1 00:11:41.377 00:11:41.377 ' 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:11:41.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:41.377 --rc genhtml_branch_coverage=1 00:11:41.377 --rc genhtml_function_coverage=1 00:11:41.377 --rc genhtml_legend=1 00:11:41.377 --rc geninfo_all_blocks=1 00:11:41.377 --rc geninfo_unexecuted_blocks=1 00:11:41.377 00:11:41.377 ' 00:11:41.377 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:11:41.377 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:41.377 --rc genhtml_branch_coverage=1 00:11:41.377 --rc genhtml_function_coverage=1 00:11:41.377 --rc genhtml_legend=1 00:11:41.377 --rc geninfo_all_blocks=1 00:11:41.377 --rc geninfo_unexecuted_blocks=1 00:11:41.378 00:11:41.378 ' 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:11:41.378 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:41.378 --rc genhtml_branch_coverage=1 00:11:41.378 --rc genhtml_function_coverage=1 00:11:41.378 --rc genhtml_legend=1 00:11:41.378 --rc geninfo_all_blocks=1 00:11:41.378 --rc geninfo_unexecuted_blocks=1 00:11:41.378 00:11:41.378 ' 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # uname -s 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@15 -- # shopt -s extglob 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@5 -- # export PATH 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@51 -- # : 0 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:11:41.378 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@55 -- # have_pci_nics=0 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@11 -- # MALLOC_BDEV_SIZE=64 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@14 -- # nvmftestinit 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@474 -- # prepare_net_devs 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@436 -- # local -g is_hw=no 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@438 -- # remove_spdk_ns 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@309 -- # xtrace_disable 00:11:41.378 08:47:31 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # pci_devs=() 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # local -a pci_devs 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # pci_net_devs=() 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # pci_drivers=() 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # local -A pci_drivers 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # net_devs=() 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # local -ga net_devs 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # e810=() 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # local -ga e810 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # x722=() 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # local -ga x722 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # mlx=() 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # local -ga mlx 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:11:49.518 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:49.518 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:11:49.519 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ up == up ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:11:49.519 Found net devices under 0000:4b:00.0: cvl_0_0 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ up == up ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:11:49.519 Found net devices under 0000:4b:00.1: cvl_0_1 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # is_hw=yes 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:11:49.519 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:11:49.519 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.636 ms 00:11:49.519 00:11:49.519 --- 10.0.0.2 ping statistics --- 00:11:49.519 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:49.519 rtt min/avg/max/mdev = 0.636/0.636/0.636/0.000 ms 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:11:49.519 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:11:49.519 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.307 ms 00:11:49.519 00:11:49.519 --- 10.0.0.1 ping statistics --- 00:11:49.519 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:11:49.519 rtt min/avg/max/mdev = 0.307/0.307/0.307/0.000 ms 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@448 -- # return 0 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@15 -- # nvmfappstart -m 0xF --wait-for-rpc 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@724 -- # xtrace_disable 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@507 -- # nvmfpid=1721062 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@508 -- # waitforlisten 1721062 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@831 -- # '[' -z 1721062 ']' 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@836 -- # local max_retries=100 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:49.519 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@840 -- # xtrace_disable 00:11:49.519 08:47:38 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.519 [2024-11-06 08:47:38.728560] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:11:49.519 [2024-11-06 08:47:38.728619] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:49.519 [2024-11-06 08:47:38.810087] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:49.519 [2024-11-06 08:47:38.849969] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:49.519 [2024-11-06 08:47:38.850005] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:49.519 [2024-11-06 08:47:38.850014] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:49.519 [2024-11-06 08:47:38.850020] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:49.519 [2024-11-06 08:47:38.850026] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:49.519 [2024-11-06 08:47:38.851596] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:11:49.519 [2024-11-06 08:47:38.851710] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:11:49.519 [2024-11-06 08:47:38.851870] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:11:49.519 [2024-11-06 08:47:38.851871] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:11:49.519 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:11:49.519 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@864 -- # return 0 00:11:49.519 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:11:49.519 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@730 -- # xtrace_disable 00:11:49.519 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.519 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:11:49.520 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@18 -- # rpc_cmd bdev_set_options -p 5 -c 1 00:11:49.520 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:49.520 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.520 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.520 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@19 -- # rpc_cmd framework_start_init 00:11:49.520 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:49.520 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.520 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.520 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:11:49.520 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.781 [2024-11-06 08:47:39.636199] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.781 Malloc0 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:49.781 [2024-11-06 08:47:39.695420] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@28 -- # WRITE_PID=1721136 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@30 -- # READ_PID=1721140 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x10 -i 1 --json /dev/fd/63 -q 128 -o 4096 -w write -t 1 -s 256 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # gen_nvmf_target_json 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:11:49.781 { 00:11:49.781 "params": { 00:11:49.781 "name": "Nvme$subsystem", 00:11:49.781 "trtype": "$TEST_TRANSPORT", 00:11:49.781 "traddr": "$NVMF_FIRST_TARGET_IP", 00:11:49.781 "adrfam": "ipv4", 00:11:49.781 "trsvcid": "$NVMF_PORT", 00:11:49.781 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:11:49.781 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:11:49.781 "hdgst": ${hdgst:-false}, 00:11:49.781 "ddgst": ${ddgst:-false} 00:11:49.781 }, 00:11:49.781 "method": "bdev_nvme_attach_controller" 00:11:49.781 } 00:11:49.781 EOF 00:11:49.781 )") 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@32 -- # FLUSH_PID=1721143 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x20 -i 2 --json /dev/fd/63 -q 128 -o 4096 -w read -t 1 -s 256 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # gen_nvmf_target_json 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:11:49.781 { 00:11:49.781 "params": { 00:11:49.781 "name": "Nvme$subsystem", 00:11:49.781 "trtype": "$TEST_TRANSPORT", 00:11:49.781 "traddr": "$NVMF_FIRST_TARGET_IP", 00:11:49.781 "adrfam": "ipv4", 00:11:49.781 "trsvcid": "$NVMF_PORT", 00:11:49.781 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:11:49.781 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:11:49.781 "hdgst": ${hdgst:-false}, 00:11:49.781 "ddgst": ${ddgst:-false} 00:11:49.781 }, 00:11:49.781 "method": "bdev_nvme_attach_controller" 00:11:49.781 } 00:11:49.781 EOF 00:11:49.781 )") 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@34 -- # UNMAP_PID=1721147 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x40 -i 3 --json /dev/fd/63 -q 128 -o 4096 -w flush -t 1 -s 256 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # gen_nvmf_target_json 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@35 -- # sync 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:11:49.781 { 00:11:49.781 "params": { 00:11:49.781 "name": "Nvme$subsystem", 00:11:49.781 "trtype": "$TEST_TRANSPORT", 00:11:49.781 "traddr": "$NVMF_FIRST_TARGET_IP", 00:11:49.781 "adrfam": "ipv4", 00:11:49.781 "trsvcid": "$NVMF_PORT", 00:11:49.781 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:11:49.781 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:11:49.781 "hdgst": ${hdgst:-false}, 00:11:49.781 "ddgst": ${ddgst:-false} 00:11:49.781 }, 00:11:49.781 "method": "bdev_nvme_attach_controller" 00:11:49.781 } 00:11:49.781 EOF 00:11:49.781 )") 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x80 -i 4 --json /dev/fd/63 -q 128 -o 4096 -w unmap -t 1 -s 256 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # gen_nvmf_target_json 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:11:49.781 { 00:11:49.781 "params": { 00:11:49.781 "name": "Nvme$subsystem", 00:11:49.781 "trtype": "$TEST_TRANSPORT", 00:11:49.781 "traddr": "$NVMF_FIRST_TARGET_IP", 00:11:49.781 "adrfam": "ipv4", 00:11:49.781 "trsvcid": "$NVMF_PORT", 00:11:49.781 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:11:49.781 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:11:49.781 "hdgst": ${hdgst:-false}, 00:11:49.781 "ddgst": ${ddgst:-false} 00:11:49.781 }, 00:11:49.781 "method": "bdev_nvme_attach_controller" 00:11:49.781 } 00:11:49.781 EOF 00:11:49.781 )") 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@37 -- # wait 1721136 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:11:49.781 "params": { 00:11:49.781 "name": "Nvme1", 00:11:49.781 "trtype": "tcp", 00:11:49.781 "traddr": "10.0.0.2", 00:11:49.781 "adrfam": "ipv4", 00:11:49.781 "trsvcid": "4420", 00:11:49.781 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:11:49.781 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:11:49.781 "hdgst": false, 00:11:49.781 "ddgst": false 00:11:49.781 }, 00:11:49.781 "method": "bdev_nvme_attach_controller" 00:11:49.781 }' 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:11:49.781 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:11:49.781 "params": { 00:11:49.781 "name": "Nvme1", 00:11:49.781 "trtype": "tcp", 00:11:49.781 "traddr": "10.0.0.2", 00:11:49.781 "adrfam": "ipv4", 00:11:49.781 "trsvcid": "4420", 00:11:49.781 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:11:49.782 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:11:49.782 "hdgst": false, 00:11:49.782 "ddgst": false 00:11:49.782 }, 00:11:49.782 "method": "bdev_nvme_attach_controller" 00:11:49.782 }' 00:11:49.782 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:11:49.782 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:11:49.782 "params": { 00:11:49.782 "name": "Nvme1", 00:11:49.782 "trtype": "tcp", 00:11:49.782 "traddr": "10.0.0.2", 00:11:49.782 "adrfam": "ipv4", 00:11:49.782 "trsvcid": "4420", 00:11:49.782 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:11:49.782 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:11:49.782 "hdgst": false, 00:11:49.782 "ddgst": false 00:11:49.782 }, 00:11:49.782 "method": "bdev_nvme_attach_controller" 00:11:49.782 }' 00:11:49.782 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:11:49.782 08:47:39 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:11:49.782 "params": { 00:11:49.782 "name": "Nvme1", 00:11:49.782 "trtype": "tcp", 00:11:49.782 "traddr": "10.0.0.2", 00:11:49.782 "adrfam": "ipv4", 00:11:49.782 "trsvcid": "4420", 00:11:49.782 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:11:49.782 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:11:49.782 "hdgst": false, 00:11:49.782 "ddgst": false 00:11:49.782 }, 00:11:49.782 "method": "bdev_nvme_attach_controller" 00:11:49.782 }' 00:11:49.782 [2024-11-06 08:47:39.750997] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:11:49.782 [2024-11-06 08:47:39.751047] [ DPDK EAL parameters: bdevperf -c 0x10 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:11:49.782 [2024-11-06 08:47:39.751719] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:11:49.782 [2024-11-06 08:47:39.751771] [ DPDK EAL parameters: bdevperf -c 0x40 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk3 --proc-type=auto ] 00:11:49.782 [2024-11-06 08:47:39.756130] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:11:49.782 [2024-11-06 08:47:39.756176] [ DPDK EAL parameters: bdevperf -c 0x80 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk4 --proc-type=auto ] 00:11:49.782 [2024-11-06 08:47:39.756204] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:11:49.782 [2024-11-06 08:47:39.756255] [ DPDK EAL parameters: bdevperf -c 0x20 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk2 --proc-type=auto ] 00:11:49.782 [2024-11-06 08:47:39.885631] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:50.042 [2024-11-06 08:47:39.914040] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:11:50.042 [2024-11-06 08:47:39.952237] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:50.042 [2024-11-06 08:47:39.981469] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:11:50.042 [2024-11-06 08:47:40.000556] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:50.042 [2024-11-06 08:47:40.030764] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:11:50.042 [2024-11-06 08:47:40.064601] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:11:50.042 [2024-11-06 08:47:40.093696] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:11:50.042 Running I/O for 1 seconds... 00:11:50.303 Running I/O for 1 seconds... 00:11:50.303 Running I/O for 1 seconds... 00:11:50.303 Running I/O for 1 seconds... 00:11:51.244 186760.00 IOPS, 729.53 MiB/s 00:11:51.244 Latency(us) 00:11:51.244 [2024-11-06T07:47:41.358Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:51.245 Job: Nvme1n1 (Core Mask 0x40, workload: flush, depth: 128, IO size: 4096) 00:11:51.245 Nvme1n1 : 1.00 186389.63 728.08 0.00 0.00 683.10 312.32 1979.73 00:11:51.245 [2024-11-06T07:47:41.358Z] =================================================================================================================== 00:11:51.245 [2024-11-06T07:47:41.358Z] Total : 186389.63 728.08 0.00 0.00 683.10 312.32 1979.73 00:11:51.245 7769.00 IOPS, 30.35 MiB/s 00:11:51.245 Latency(us) 00:11:51.245 [2024-11-06T07:47:41.358Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:51.245 Job: Nvme1n1 (Core Mask 0x10, workload: write, depth: 128, IO size: 4096) 00:11:51.245 Nvme1n1 : 1.02 7783.79 30.41 0.00 0.00 16306.08 7045.12 26869.76 00:11:51.245 [2024-11-06T07:47:41.358Z] =================================================================================================================== 00:11:51.245 [2024-11-06T07:47:41.358Z] Total : 7783.79 30.41 0.00 0.00 16306.08 7045.12 26869.76 00:11:51.245 20360.00 IOPS, 79.53 MiB/s 00:11:51.245 Latency(us) 00:11:51.245 [2024-11-06T07:47:41.358Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:51.245 Job: Nvme1n1 (Core Mask 0x80, workload: unmap, depth: 128, IO size: 4096) 00:11:51.245 Nvme1n1 : 1.01 20417.43 79.76 0.00 0.00 6253.40 3194.88 15837.87 00:11:51.245 [2024-11-06T07:47:41.358Z] =================================================================================================================== 00:11:51.245 [2024-11-06T07:47:41.358Z] Total : 20417.43 79.76 0.00 0.00 6253.40 3194.88 15837.87 00:11:51.245 7503.00 IOPS, 29.31 MiB/s 00:11:51.245 Latency(us) 00:11:51.245 [2024-11-06T07:47:41.358Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:11:51.245 Job: Nvme1n1 (Core Mask 0x20, workload: read, depth: 128, IO size: 4096) 00:11:51.245 Nvme1n1 : 1.01 7591.72 29.66 0.00 0.00 16815.99 3850.24 39758.51 00:11:51.245 [2024-11-06T07:47:41.358Z] =================================================================================================================== 00:11:51.245 [2024-11-06T07:47:41.358Z] Total : 7591.72 29.66 0.00 0.00 16815.99 3850.24 39758.51 00:11:51.245 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@38 -- # wait 1721140 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@39 -- # wait 1721143 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@40 -- # wait 1721147 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@42 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@44 -- # trap - SIGINT SIGTERM EXIT 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@46 -- # nvmftestfini 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@514 -- # nvmfcleanup 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@121 -- # sync 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@124 -- # set +e 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@125 -- # for i in {1..20} 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:11:51.505 rmmod nvme_tcp 00:11:51.505 rmmod nvme_fabrics 00:11:51.505 rmmod nvme_keyring 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@128 -- # set -e 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@129 -- # return 0 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@515 -- # '[' -n 1721062 ']' 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@516 -- # killprocess 1721062 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@950 -- # '[' -z 1721062 ']' 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@954 -- # kill -0 1721062 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@955 -- # uname 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1721062 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1721062' 00:11:51.505 killing process with pid 1721062 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@969 -- # kill 1721062 00:11:51.505 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@974 -- # wait 1721062 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@297 -- # iptr 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # iptables-save 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # iptables-restore 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@302 -- # remove_spdk_ns 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:51.765 08:47:41 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:53.740 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:11:53.740 00:11:53.740 real 0m12.683s 00:11:53.740 user 0m18.302s 00:11:53.740 sys 0m6.946s 00:11:53.740 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@1126 -- # xtrace_disable 00:11:53.740 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:11:53.740 ************************************ 00:11:53.740 END TEST nvmf_bdev_io_wait 00:11:53.740 ************************************ 00:11:53.740 08:47:43 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@30 -- # run_test nvmf_queue_depth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:11:53.740 08:47:43 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:11:53.740 08:47:43 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:11:53.740 08:47:43 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:11:53.740 ************************************ 00:11:53.740 START TEST nvmf_queue_depth 00:11:53.740 ************************************ 00:11:53.740 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:11:54.001 * Looking for test storage... 00:11:54.001 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1689 -- # lcov --version 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@333 -- # local ver1 ver1_l 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@334 -- # local ver2 ver2_l 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@336 -- # IFS=.-: 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@336 -- # read -ra ver1 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@337 -- # IFS=.-: 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@337 -- # read -ra ver2 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@338 -- # local 'op=<' 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@340 -- # ver1_l=2 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@341 -- # ver2_l=1 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@344 -- # case "$op" in 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@345 -- # : 1 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v = 0 )) 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:11:54.001 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@365 -- # decimal 1 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=1 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 1 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@365 -- # ver1[v]=1 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@366 -- # decimal 2 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=2 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 2 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@366 -- # ver2[v]=2 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@368 -- # return 0 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:11:54.002 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:54.002 --rc genhtml_branch_coverage=1 00:11:54.002 --rc genhtml_function_coverage=1 00:11:54.002 --rc genhtml_legend=1 00:11:54.002 --rc geninfo_all_blocks=1 00:11:54.002 --rc geninfo_unexecuted_blocks=1 00:11:54.002 00:11:54.002 ' 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:11:54.002 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:54.002 --rc genhtml_branch_coverage=1 00:11:54.002 --rc genhtml_function_coverage=1 00:11:54.002 --rc genhtml_legend=1 00:11:54.002 --rc geninfo_all_blocks=1 00:11:54.002 --rc geninfo_unexecuted_blocks=1 00:11:54.002 00:11:54.002 ' 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:11:54.002 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:54.002 --rc genhtml_branch_coverage=1 00:11:54.002 --rc genhtml_function_coverage=1 00:11:54.002 --rc genhtml_legend=1 00:11:54.002 --rc geninfo_all_blocks=1 00:11:54.002 --rc geninfo_unexecuted_blocks=1 00:11:54.002 00:11:54.002 ' 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:11:54.002 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:11:54.002 --rc genhtml_branch_coverage=1 00:11:54.002 --rc genhtml_function_coverage=1 00:11:54.002 --rc genhtml_legend=1 00:11:54.002 --rc geninfo_all_blocks=1 00:11:54.002 --rc geninfo_unexecuted_blocks=1 00:11:54.002 00:11:54.002 ' 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@7 -- # uname -s 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@15 -- # shopt -s extglob 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:54.002 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@5 -- # export PATH 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@51 -- # : 0 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:11:54.003 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@55 -- # have_pci_nics=0 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@14 -- # MALLOC_BDEV_SIZE=64 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@15 -- # MALLOC_BLOCK_SIZE=512 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@17 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@19 -- # nvmftestinit 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@474 -- # prepare_net_devs 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@436 -- # local -g is_hw=no 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@438 -- # remove_spdk_ns 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:11:54.003 08:47:43 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:54.003 08:47:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:11:54.003 08:47:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:11:54.003 08:47:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@309 -- # xtrace_disable 00:11:54.003 08:47:44 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@315 -- # pci_devs=() 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@319 -- # net_devs=() 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@320 -- # e810=() 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@320 -- # local -ga e810 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@321 -- # x722=() 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@321 -- # local -ga x722 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@322 -- # mlx=() 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@322 -- # local -ga mlx 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:02.151 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:12:02.152 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:12:02.152 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:12:02.152 Found net devices under 0000:4b:00.0: cvl_0_0 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:12:02.152 Found net devices under 0000:4b:00.1: cvl_0_1 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@440 -- # is_hw=yes 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:02.152 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:02.152 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.553 ms 00:12:02.152 00:12:02.152 --- 10.0.0.2 ping statistics --- 00:12:02.152 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:02.152 rtt min/avg/max/mdev = 0.553/0.553/0.553/0.000 ms 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:02.152 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:02.152 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.283 ms 00:12:02.152 00:12:02.152 --- 10.0.0.1 ping statistics --- 00:12:02.152 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:02.152 rtt min/avg/max/mdev = 0.283/0.283/0.283/0.000 ms 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@448 -- # return 0 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@21 -- # nvmfappstart -m 0x2 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@724 -- # xtrace_disable 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@507 -- # nvmfpid=1725800 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@508 -- # waitforlisten 1725800 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@831 -- # '[' -z 1725800 ']' 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:02.152 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:02.152 08:47:51 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:02.152 [2024-11-06 08:47:51.485695] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:12:02.152 [2024-11-06 08:47:51.485767] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:02.152 [2024-11-06 08:47:51.590010] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:02.152 [2024-11-06 08:47:51.640366] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:02.152 [2024-11-06 08:47:51.640417] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:02.152 [2024-11-06 08:47:51.640426] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:02.153 [2024-11-06 08:47:51.640434] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:02.153 [2024-11-06 08:47:51.640440] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:02.153 [2024-11-06 08:47:51.641215] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@864 -- # return 0 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@730 -- # xtrace_disable 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:02.415 [2024-11-06 08:47:52.348084] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@24 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:02.415 Malloc0 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:02.415 [2024-11-06 08:47:52.409271] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@30 -- # bdevperf_pid=1726112 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 1024 -o 4096 -w verify -t 10 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@32 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@33 -- # waitforlisten 1726112 /var/tmp/bdevperf.sock 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@831 -- # '[' -z 1726112 ']' 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:12:02.415 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:02.415 08:47:52 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:02.415 [2024-11-06 08:47:52.468931] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:12:02.415 [2024-11-06 08:47:52.468996] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1726112 ] 00:12:02.676 [2024-11-06 08:47:52.544336] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:02.676 [2024-11-06 08:47:52.586209] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:03.247 08:47:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:03.247 08:47:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@864 -- # return 0 00:12:03.247 08:47:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@34 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:12:03.247 08:47:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:03.247 08:47:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:03.509 NVMe0n1 00:12:03.509 08:47:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:03.509 08:47:53 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:12:03.509 Running I/O for 10 seconds... 00:12:05.394 10240.00 IOPS, 40.00 MiB/s [2024-11-06T07:47:56.891Z] 11085.00 IOPS, 43.30 MiB/s [2024-11-06T07:47:57.831Z] 11264.00 IOPS, 44.00 MiB/s [2024-11-06T07:47:58.768Z] 11388.00 IOPS, 44.48 MiB/s [2024-11-06T07:47:59.708Z] 11470.40 IOPS, 44.81 MiB/s [2024-11-06T07:48:00.647Z] 11465.67 IOPS, 44.79 MiB/s [2024-11-06T07:48:01.586Z] 11539.43 IOPS, 45.08 MiB/s [2024-11-06T07:48:02.527Z] 11532.88 IOPS, 45.05 MiB/s [2024-11-06T07:48:03.912Z] 11583.44 IOPS, 45.25 MiB/s [2024-11-06T07:48:03.912Z] 11570.90 IOPS, 45.20 MiB/s 00:12:13.799 Latency(us) 00:12:13.799 [2024-11-06T07:48:03.912Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:13.799 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 1024, IO size: 4096) 00:12:13.799 Verification LBA range: start 0x0 length 0x4000 00:12:13.799 NVMe0n1 : 10.06 11607.06 45.34 0.00 0.00 87914.16 25012.91 75147.95 00:12:13.799 [2024-11-06T07:48:03.912Z] =================================================================================================================== 00:12:13.799 [2024-11-06T07:48:03.912Z] Total : 11607.06 45.34 0.00 0.00 87914.16 25012.91 75147.95 00:12:13.799 { 00:12:13.799 "results": [ 00:12:13.799 { 00:12:13.799 "job": "NVMe0n1", 00:12:13.799 "core_mask": "0x1", 00:12:13.799 "workload": "verify", 00:12:13.799 "status": "finished", 00:12:13.799 "verify_range": { 00:12:13.799 "start": 0, 00:12:13.799 "length": 16384 00:12:13.799 }, 00:12:13.799 "queue_depth": 1024, 00:12:13.799 "io_size": 4096, 00:12:13.799 "runtime": 10.057067, 00:12:13.799 "iops": 11607.061979402146, 00:12:13.799 "mibps": 45.34008585703963, 00:12:13.799 "io_failed": 0, 00:12:13.799 "io_timeout": 0, 00:12:13.799 "avg_latency_us": 87914.16106156785, 00:12:13.799 "min_latency_us": 25012.906666666666, 00:12:13.799 "max_latency_us": 75147.94666666667 00:12:13.799 } 00:12:13.799 ], 00:12:13.799 "core_count": 1 00:12:13.799 } 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@39 -- # killprocess 1726112 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@950 -- # '[' -z 1726112 ']' 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@954 -- # kill -0 1726112 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@955 -- # uname 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1726112 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1726112' 00:12:13.799 killing process with pid 1726112 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@969 -- # kill 1726112 00:12:13.799 Received shutdown signal, test time was about 10.000000 seconds 00:12:13.799 00:12:13.799 Latency(us) 00:12:13.799 [2024-11-06T07:48:03.912Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:13.799 [2024-11-06T07:48:03.912Z] =================================================================================================================== 00:12:13.799 [2024-11-06T07:48:03.912Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@974 -- # wait 1726112 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- target/queue_depth.sh@43 -- # nvmftestfini 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@514 -- # nvmfcleanup 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@121 -- # sync 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@124 -- # set +e 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:13.799 rmmod nvme_tcp 00:12:13.799 rmmod nvme_fabrics 00:12:13.799 rmmod nvme_keyring 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@128 -- # set -e 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@129 -- # return 0 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@515 -- # '[' -n 1725800 ']' 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@516 -- # killprocess 1725800 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@950 -- # '[' -z 1725800 ']' 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@954 -- # kill -0 1725800 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@955 -- # uname 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:13.799 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1725800 00:12:14.060 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:12:14.060 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:12:14.060 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1725800' 00:12:14.060 killing process with pid 1725800 00:12:14.060 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@969 -- # kill 1725800 00:12:14.060 08:48:03 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@974 -- # wait 1725800 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@297 -- # iptr 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@789 -- # iptables-save 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@789 -- # iptables-restore 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:14.060 08:48:04 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:16.609 00:12:16.609 real 0m22.325s 00:12:16.609 user 0m25.696s 00:12:16.609 sys 0m6.880s 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:12:16.609 ************************************ 00:12:16.609 END TEST nvmf_queue_depth 00:12:16.609 ************************************ 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@31 -- # run_test nvmf_target_multipath /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:12:16.609 ************************************ 00:12:16.609 START TEST nvmf_target_multipath 00:12:16.609 ************************************ 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:12:16.609 * Looking for test storage... 00:12:16.609 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1689 -- # lcov --version 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@336 -- # IFS=.-: 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@336 -- # read -ra ver1 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@337 -- # IFS=.-: 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@337 -- # read -ra ver2 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@338 -- # local 'op=<' 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@340 -- # ver1_l=2 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@341 -- # ver2_l=1 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@344 -- # case "$op" in 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@345 -- # : 1 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@365 -- # decimal 1 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=1 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 1 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@365 -- # ver1[v]=1 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@366 -- # decimal 2 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=2 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 2 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@366 -- # ver2[v]=2 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@368 -- # return 0 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:16.609 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:12:16.609 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:16.609 --rc genhtml_branch_coverage=1 00:12:16.609 --rc genhtml_function_coverage=1 00:12:16.609 --rc genhtml_legend=1 00:12:16.609 --rc geninfo_all_blocks=1 00:12:16.609 --rc geninfo_unexecuted_blocks=1 00:12:16.609 00:12:16.609 ' 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:12:16.610 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:16.610 --rc genhtml_branch_coverage=1 00:12:16.610 --rc genhtml_function_coverage=1 00:12:16.610 --rc genhtml_legend=1 00:12:16.610 --rc geninfo_all_blocks=1 00:12:16.610 --rc geninfo_unexecuted_blocks=1 00:12:16.610 00:12:16.610 ' 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:12:16.610 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:16.610 --rc genhtml_branch_coverage=1 00:12:16.610 --rc genhtml_function_coverage=1 00:12:16.610 --rc genhtml_legend=1 00:12:16.610 --rc geninfo_all_blocks=1 00:12:16.610 --rc geninfo_unexecuted_blocks=1 00:12:16.610 00:12:16.610 ' 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:12:16.610 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:16.610 --rc genhtml_branch_coverage=1 00:12:16.610 --rc genhtml_function_coverage=1 00:12:16.610 --rc genhtml_legend=1 00:12:16.610 --rc geninfo_all_blocks=1 00:12:16.610 --rc geninfo_unexecuted_blocks=1 00:12:16.610 00:12:16.610 ' 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@7 -- # uname -s 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@15 -- # shopt -s extglob 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@5 -- # export PATH 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@51 -- # : 0 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:16.610 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@11 -- # MALLOC_BDEV_SIZE=64 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@43 -- # nvmftestinit 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@474 -- # prepare_net_devs 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@436 -- # local -g is_hw=no 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@438 -- # remove_spdk_ns 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@309 -- # xtrace_disable 00:12:16.610 08:48:06 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@315 -- # pci_devs=() 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@319 -- # net_devs=() 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@320 -- # e810=() 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@320 -- # local -ga e810 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@321 -- # x722=() 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@321 -- # local -ga x722 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@322 -- # mlx=() 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@322 -- # local -ga mlx 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:12:24.752 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:12:24.752 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:24.752 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:12:24.753 Found net devices under 0000:4b:00.0: cvl_0_0 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:12:24.753 Found net devices under 0000:4b:00.1: cvl_0_1 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@440 -- # is_hw=yes 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:24.753 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:24.753 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.583 ms 00:12:24.753 00:12:24.753 --- 10.0.0.2 ping statistics --- 00:12:24.753 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:24.753 rtt min/avg/max/mdev = 0.583/0.583/0.583/0.000 ms 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:24.753 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:24.753 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.269 ms 00:12:24.753 00:12:24.753 --- 10.0.0.1 ping statistics --- 00:12:24.753 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:24.753 rtt min/avg/max/mdev = 0.269/0.269/0.269/0.000 ms 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@448 -- # return 0 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@45 -- # '[' -z ']' 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@46 -- # echo 'only one NIC for nvmf test' 00:12:24.753 only one NIC for nvmf test 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@47 -- # nvmftestfini 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@514 -- # nvmfcleanup 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:24.753 rmmod nvme_tcp 00:12:24.753 rmmod nvme_fabrics 00:12:24.753 rmmod nvme_keyring 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-save 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:12:24.753 08:48:13 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-restore 00:12:24.753 08:48:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:24.753 08:48:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:24.753 08:48:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:24.753 08:48:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:24.753 08:48:14 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@48 -- # exit 0 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- target/multipath.sh@1 -- # nvmftestfini 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@514 -- # nvmfcleanup 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-save 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-restore 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:26.136 00:12:26.136 real 0m9.929s 00:12:26.136 user 0m2.118s 00:12:26.136 sys 0m5.744s 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:12:26.136 ************************************ 00:12:26.136 END TEST nvmf_target_multipath 00:12:26.136 ************************************ 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@32 -- # run_test nvmf_zcopy /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:12:26.136 ************************************ 00:12:26.136 START TEST nvmf_zcopy 00:12:26.136 ************************************ 00:12:26.136 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:12:26.398 * Looking for test storage... 00:12:26.398 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1689 -- # lcov --version 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@336 -- # IFS=.-: 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@336 -- # read -ra ver1 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@337 -- # IFS=.-: 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@337 -- # read -ra ver2 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@338 -- # local 'op=<' 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@340 -- # ver1_l=2 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@341 -- # ver2_l=1 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@344 -- # case "$op" in 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@345 -- # : 1 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@365 -- # decimal 1 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@353 -- # local d=1 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@355 -- # echo 1 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@365 -- # ver1[v]=1 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@366 -- # decimal 2 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@353 -- # local d=2 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@355 -- # echo 2 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@366 -- # ver2[v]=2 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@368 -- # return 0 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:12:26.398 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:26.398 --rc genhtml_branch_coverage=1 00:12:26.398 --rc genhtml_function_coverage=1 00:12:26.398 --rc genhtml_legend=1 00:12:26.398 --rc geninfo_all_blocks=1 00:12:26.398 --rc geninfo_unexecuted_blocks=1 00:12:26.398 00:12:26.398 ' 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:12:26.398 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:26.398 --rc genhtml_branch_coverage=1 00:12:26.398 --rc genhtml_function_coverage=1 00:12:26.398 --rc genhtml_legend=1 00:12:26.398 --rc geninfo_all_blocks=1 00:12:26.398 --rc geninfo_unexecuted_blocks=1 00:12:26.398 00:12:26.398 ' 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:12:26.398 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:26.398 --rc genhtml_branch_coverage=1 00:12:26.398 --rc genhtml_function_coverage=1 00:12:26.398 --rc genhtml_legend=1 00:12:26.398 --rc geninfo_all_blocks=1 00:12:26.398 --rc geninfo_unexecuted_blocks=1 00:12:26.398 00:12:26.398 ' 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:12:26.398 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:26.398 --rc genhtml_branch_coverage=1 00:12:26.398 --rc genhtml_function_coverage=1 00:12:26.398 --rc genhtml_legend=1 00:12:26.398 --rc geninfo_all_blocks=1 00:12:26.398 --rc geninfo_unexecuted_blocks=1 00:12:26.398 00:12:26.398 ' 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@7 -- # uname -s 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:26.398 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@15 -- # shopt -s extglob 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@5 -- # export PATH 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@51 -- # : 0 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:26.399 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@12 -- # nvmftestinit 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@474 -- # prepare_net_devs 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@436 -- # local -g is_hw=no 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@438 -- # remove_spdk_ns 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@309 -- # xtrace_disable 00:12:26.399 08:48:16 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@315 -- # pci_devs=() 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@315 -- # local -a pci_devs 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@316 -- # pci_net_devs=() 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@317 -- # pci_drivers=() 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@317 -- # local -A pci_drivers 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@319 -- # net_devs=() 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@319 -- # local -ga net_devs 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@320 -- # e810=() 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@320 -- # local -ga e810 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@321 -- # x722=() 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@321 -- # local -ga x722 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@322 -- # mlx=() 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@322 -- # local -ga mlx 00:12:34.538 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:12:34.539 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:12:34.539 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:12:34.539 Found net devices under 0000:4b:00.0: cvl_0_0 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ up == up ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:12:34.539 Found net devices under 0000:4b:00.1: cvl_0_1 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@440 -- # is_hw=yes 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:12:34.539 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:34.539 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.691 ms 00:12:34.539 00:12:34.539 --- 10.0.0.2 ping statistics --- 00:12:34.539 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:34.539 rtt min/avg/max/mdev = 0.691/0.691/0.691/0.000 ms 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:34.539 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:34.539 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.325 ms 00:12:34.539 00:12:34.539 --- 10.0.0.1 ping statistics --- 00:12:34.539 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:34.539 rtt min/avg/max/mdev = 0.325/0.325/0.325/0.000 ms 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@448 -- # return 0 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@13 -- # nvmfappstart -m 0x2 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@724 -- # xtrace_disable 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@507 -- # nvmfpid=1737412 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@508 -- # waitforlisten 1737412 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@831 -- # '[' -z 1737412 ']' 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:34.539 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@836 -- # local max_retries=100 00:12:34.540 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:34.540 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:34.540 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@840 -- # xtrace_disable 00:12:34.540 08:48:23 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:34.540 [2024-11-06 08:48:23.834353] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:12:34.540 [2024-11-06 08:48:23.834421] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:34.540 [2024-11-06 08:48:23.935303] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:34.540 [2024-11-06 08:48:23.988637] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:34.540 [2024-11-06 08:48:23.988696] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:34.540 [2024-11-06 08:48:23.988705] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:34.540 [2024-11-06 08:48:23.988713] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:34.540 [2024-11-06 08:48:23.988719] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:34.540 [2024-11-06 08:48:23.989490] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:12:34.540 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:12:34.540 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@864 -- # return 0 00:12:34.540 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:12:34.540 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@730 -- # xtrace_disable 00:12:34.540 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@15 -- # '[' tcp '!=' tcp ']' 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@22 -- # rpc_cmd nvmf_create_transport -t tcp -o -c 0 --zcopy 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:34.801 [2024-11-06 08:48:24.694429] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:34.801 [2024-11-06 08:48:24.718734] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@29 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc0 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:34.801 malloc0 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -t 10 -q 128 -w verify -o 8192 00:12:34.801 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@33 -- # gen_nvmf_target_json 00:12:34.802 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@558 -- # config=() 00:12:34.802 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@558 -- # local subsystem config 00:12:34.802 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:12:34.802 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:12:34.802 { 00:12:34.802 "params": { 00:12:34.802 "name": "Nvme$subsystem", 00:12:34.802 "trtype": "$TEST_TRANSPORT", 00:12:34.802 "traddr": "$NVMF_FIRST_TARGET_IP", 00:12:34.802 "adrfam": "ipv4", 00:12:34.802 "trsvcid": "$NVMF_PORT", 00:12:34.802 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:12:34.802 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:12:34.802 "hdgst": ${hdgst:-false}, 00:12:34.802 "ddgst": ${ddgst:-false} 00:12:34.802 }, 00:12:34.802 "method": "bdev_nvme_attach_controller" 00:12:34.802 } 00:12:34.802 EOF 00:12:34.802 )") 00:12:34.802 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@580 -- # cat 00:12:34.802 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # jq . 00:12:34.802 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@583 -- # IFS=, 00:12:34.802 08:48:24 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:12:34.802 "params": { 00:12:34.802 "name": "Nvme1", 00:12:34.802 "trtype": "tcp", 00:12:34.802 "traddr": "10.0.0.2", 00:12:34.802 "adrfam": "ipv4", 00:12:34.802 "trsvcid": "4420", 00:12:34.802 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:12:34.802 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:12:34.802 "hdgst": false, 00:12:34.802 "ddgst": false 00:12:34.802 }, 00:12:34.802 "method": "bdev_nvme_attach_controller" 00:12:34.802 }' 00:12:34.802 [2024-11-06 08:48:24.819147] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:12:34.802 [2024-11-06 08:48:24.819212] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1737473 ] 00:12:34.802 [2024-11-06 08:48:24.894478] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:35.062 [2024-11-06 08:48:24.937045] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:35.062 Running I/O for 10 seconds... 00:12:37.392 6587.00 IOPS, 51.46 MiB/s [2024-11-06T07:48:28.447Z] 7251.50 IOPS, 56.65 MiB/s [2024-11-06T07:48:29.389Z] 8061.67 IOPS, 62.98 MiB/s [2024-11-06T07:48:30.331Z] 8474.00 IOPS, 66.20 MiB/s [2024-11-06T07:48:31.380Z] 8721.80 IOPS, 68.14 MiB/s [2024-11-06T07:48:32.321Z] 8887.33 IOPS, 69.43 MiB/s [2024-11-06T07:48:33.261Z] 9006.29 IOPS, 70.36 MiB/s [2024-11-06T07:48:34.200Z] 9096.38 IOPS, 71.07 MiB/s [2024-11-06T07:48:35.584Z] 9165.56 IOPS, 71.61 MiB/s [2024-11-06T07:48:35.584Z] 9221.00 IOPS, 72.04 MiB/s 00:12:45.471 Latency(us) 00:12:45.471 [2024-11-06T07:48:35.584Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:45.471 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 8192) 00:12:45.471 Verification LBA range: start 0x0 length 0x1000 00:12:45.471 Nvme1n1 : 10.05 9184.31 71.75 0.00 0.00 13834.87 2048.00 42816.85 00:12:45.471 [2024-11-06T07:48:35.584Z] =================================================================================================================== 00:12:45.471 [2024-11-06T07:48:35.584Z] Total : 9184.31 71.75 0.00 0.00 13834.87 2048.00 42816.85 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@39 -- # perfpid=1739687 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@41 -- # xtrace_disable 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -t 5 -q 128 -w randrw -M 50 -o 8192 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@37 -- # gen_nvmf_target_json 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@558 -- # config=() 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@558 -- # local subsystem config 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:12:45.471 { 00:12:45.471 "params": { 00:12:45.471 "name": "Nvme$subsystem", 00:12:45.471 "trtype": "$TEST_TRANSPORT", 00:12:45.471 "traddr": "$NVMF_FIRST_TARGET_IP", 00:12:45.471 "adrfam": "ipv4", 00:12:45.471 "trsvcid": "$NVMF_PORT", 00:12:45.471 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:12:45.471 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:12:45.471 "hdgst": ${hdgst:-false}, 00:12:45.471 "ddgst": ${ddgst:-false} 00:12:45.471 }, 00:12:45.471 "method": "bdev_nvme_attach_controller" 00:12:45.471 } 00:12:45.471 EOF 00:12:45.471 )") 00:12:45.471 [2024-11-06 08:48:35.322967] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.322998] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@580 -- # cat 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@582 -- # jq . 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@583 -- # IFS=, 00:12:45.471 08:48:35 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:12:45.471 "params": { 00:12:45.471 "name": "Nvme1", 00:12:45.471 "trtype": "tcp", 00:12:45.471 "traddr": "10.0.0.2", 00:12:45.471 "adrfam": "ipv4", 00:12:45.471 "trsvcid": "4420", 00:12:45.471 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:12:45.471 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:12:45.471 "hdgst": false, 00:12:45.471 "ddgst": false 00:12:45.471 }, 00:12:45.471 "method": "bdev_nvme_attach_controller" 00:12:45.471 }' 00:12:45.471 [2024-11-06 08:48:35.334967] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.334976] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.346997] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.347005] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.359027] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.359035] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.371058] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.371066] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.375285] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:12:45.471 [2024-11-06 08:48:35.375337] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1739687 ] 00:12:45.471 [2024-11-06 08:48:35.383088] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.383096] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.395119] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.395127] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.407151] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.407160] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.419179] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.419187] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.431211] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.431218] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.443242] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.443249] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.445825] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:45.471 [2024-11-06 08:48:35.455272] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.455281] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.471 [2024-11-06 08:48:35.467303] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.471 [2024-11-06 08:48:35.467312] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.479334] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.479350] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.482204] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:12:45.472 [2024-11-06 08:48:35.491364] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.491372] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.503398] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.503413] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.515427] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.515438] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.527466] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.527475] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.535479] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.535487] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.543498] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.543505] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.551520] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.551527] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.559722] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.559738] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.567737] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.567748] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.472 [2024-11-06 08:48:35.575759] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.472 [2024-11-06 08:48:35.575767] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.583780] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.583790] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.591798] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.591805] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.599818] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.599825] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.607840] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.607847] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.615861] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.615868] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.623883] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.623890] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.631918] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.631927] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.639939] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.639949] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.647959] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.647968] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.655982] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.655994] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.664002] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.664015] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 Running I/O for 5 seconds... 00:12:45.732 [2024-11-06 08:48:35.672020] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.672026] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.684251] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.684266] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.690859] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.690875] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.700779] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.700795] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.709014] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.709029] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.718048] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.718063] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.726182] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.726197] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.734665] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.734680] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.743695] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.743711] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.732 [2024-11-06 08:48:35.752812] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.732 [2024-11-06 08:48:35.752827] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.733 [2024-11-06 08:48:35.761941] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.733 [2024-11-06 08:48:35.761956] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.733 [2024-11-06 08:48:35.770986] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.733 [2024-11-06 08:48:35.771001] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.733 [2024-11-06 08:48:35.779376] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.733 [2024-11-06 08:48:35.779391] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.733 [2024-11-06 08:48:35.788576] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.733 [2024-11-06 08:48:35.788591] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.733 [2024-11-06 08:48:35.797843] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.733 [2024-11-06 08:48:35.797858] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.733 [2024-11-06 08:48:35.806450] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.733 [2024-11-06 08:48:35.806468] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.733 [2024-11-06 08:48:35.815610] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.733 [2024-11-06 08:48:35.815625] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.733 [2024-11-06 08:48:35.824140] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.733 [2024-11-06 08:48:35.824155] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.733 [2024-11-06 08:48:35.833313] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.733 [2024-11-06 08:48:35.833328] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.733 [2024-11-06 08:48:35.841762] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.733 [2024-11-06 08:48:35.841778] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.851010] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.851024] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.859438] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.859452] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.868676] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.868691] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.877160] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.877174] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.886424] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.886439] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.895331] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.895346] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.903846] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.903861] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.912865] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.912880] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.921653] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.921668] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.930315] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.930330] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.938335] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.938349] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.947388] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.947403] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.956384] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.956399] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.965615] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.965630] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.974653] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.974672] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.983555] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.983569] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:35.992079] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:35.992093] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.001090] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.001106] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.009500] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.009515] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.018662] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.018677] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.027183] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.027198] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.036366] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.036381] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.044848] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.044863] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.053556] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.053570] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.061978] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.061992] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.070723] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.070737] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.079620] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.079636] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.088275] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.088290] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.097205] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.097219] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:45.994 [2024-11-06 08:48:36.106112] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:45.994 [2024-11-06 08:48:36.106127] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.115256] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.115271] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.123293] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.123307] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.131578] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.131593] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.140205] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.140226] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.148284] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.148298] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.156791] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.156805] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.165078] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.165092] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.173904] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.173919] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.182794] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.182809] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.191388] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.191402] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.200141] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.200155] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.208653] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.208667] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.217318] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.217332] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.225944] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.225959] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.234205] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.234219] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.243018] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.243032] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.251637] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.251651] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.260442] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.260456] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.269166] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.255 [2024-11-06 08:48:36.269180] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.255 [2024-11-06 08:48:36.277986] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.278001] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.256 [2024-11-06 08:48:36.287205] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.287220] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.256 [2024-11-06 08:48:36.295456] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.295471] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.256 [2024-11-06 08:48:36.304372] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.304387] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.256 [2024-11-06 08:48:36.312824] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.312839] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.256 [2024-11-06 08:48:36.321250] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.321265] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.256 [2024-11-06 08:48:36.330480] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.330494] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.256 [2024-11-06 08:48:36.339071] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.339085] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.256 [2024-11-06 08:48:36.348015] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.348029] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.256 [2024-11-06 08:48:36.357207] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.357222] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.256 [2024-11-06 08:48:36.365715] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.256 [2024-11-06 08:48:36.365730] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.374239] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.374254] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.383013] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.383028] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.391766] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.391780] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.401024] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.401038] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.409594] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.409608] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.418482] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.418497] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.426916] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.426931] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.435842] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.435857] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.444110] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.444124] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.453500] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.453514] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.462209] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.462224] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.471213] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.471227] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.480001] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.480016] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.488650] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.488665] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.497201] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.497215] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.505924] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.505938] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.514743] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.514762] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.523127] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.523141] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.532322] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.532336] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.540192] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.540206] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.549112] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.549127] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.557932] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.557947] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.566441] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.566456] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.575359] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.575374] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.516 [2024-11-06 08:48:36.583868] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.516 [2024-11-06 08:48:36.583882] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.517 [2024-11-06 08:48:36.592687] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.517 [2024-11-06 08:48:36.592702] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.517 [2024-11-06 08:48:36.601213] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.517 [2024-11-06 08:48:36.601228] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.517 [2024-11-06 08:48:36.609939] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.517 [2024-11-06 08:48:36.609953] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.517 [2024-11-06 08:48:36.619032] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.517 [2024-11-06 08:48:36.619047] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.517 [2024-11-06 08:48:36.627543] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.517 [2024-11-06 08:48:36.627559] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.777 [2024-11-06 08:48:36.636109] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.777 [2024-11-06 08:48:36.636124] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.777 [2024-11-06 08:48:36.644663] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.777 [2024-11-06 08:48:36.644678] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.777 [2024-11-06 08:48:36.653387] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.777 [2024-11-06 08:48:36.653402] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.777 [2024-11-06 08:48:36.662512] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.777 [2024-11-06 08:48:36.662527] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.777 [2024-11-06 08:48:36.671294] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.671309] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 18798.00 IOPS, 146.86 MiB/s [2024-11-06T07:48:36.891Z] [2024-11-06 08:48:36.678240] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.678255] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.688893] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.688908] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.696779] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.696793] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.705733] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.705753] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.714178] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.714193] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.722991] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.723006] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.732090] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.732105] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.740381] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.740396] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.749124] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.749139] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.757877] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.757892] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.766942] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.766957] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.775975] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.775989] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.784513] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.784527] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.793514] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.793532] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.801873] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.801888] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.810203] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.810218] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.818939] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.818954] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.827996] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.828010] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.836674] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.836689] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.845695] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.845710] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.854536] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.854552] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.863999] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.864014] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.871898] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.871913] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:46.778 [2024-11-06 08:48:36.881179] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:46.778 [2024-11-06 08:48:36.881193] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.039 [2024-11-06 08:48:36.890327] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.039 [2024-11-06 08:48:36.890343] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.039 [2024-11-06 08:48:36.898846] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.039 [2024-11-06 08:48:36.898860] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.039 [2024-11-06 08:48:36.907665] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.039 [2024-11-06 08:48:36.907679] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.039 [2024-11-06 08:48:36.916483] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.039 [2024-11-06 08:48:36.916498] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.039 [2024-11-06 08:48:36.924812] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.039 [2024-11-06 08:48:36.924827] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.039 [2024-11-06 08:48:36.934029] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.039 [2024-11-06 08:48:36.934044] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.039 [2024-11-06 08:48:36.942306] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.039 [2024-11-06 08:48:36.942322] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.039 [2024-11-06 08:48:36.951345] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:36.951361] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:36.959876] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:36.959896] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:36.968952] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:36.968967] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:36.977604] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:36.977619] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:36.985931] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:36.985945] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:36.994991] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:36.995005] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.003920] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.003935] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.012879] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.012894] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.021559] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.021574] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.030519] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.030534] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.039319] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.039334] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.048165] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.048180] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.057244] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.057259] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.065435] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.065450] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.074345] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.074360] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.083411] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.083427] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.091897] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.091911] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.100437] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.100452] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.109054] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.109070] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.117841] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.117856] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.126451] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.126469] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.135114] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.135129] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.040 [2024-11-06 08:48:37.143700] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.040 [2024-11-06 08:48:37.143715] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.152501] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.152516] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.161237] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.161251] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.170126] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.170141] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.179386] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.179401] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.188693] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.188709] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.197382] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.197396] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.206280] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.206295] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.215495] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.215510] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.224200] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.224215] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.233395] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.233409] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.242783] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.242799] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.251282] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.251296] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.260059] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.260073] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.268886] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.268901] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.277315] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.277329] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.285798] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.300 [2024-11-06 08:48:37.285812] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.300 [2024-11-06 08:48:37.294991] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.295009] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.303596] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.303610] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.312512] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.312527] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.320959] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.320974] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.329943] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.329957] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.338789] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.338803] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.346822] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.346836] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.355824] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.355838] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.365078] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.365092] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.374304] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.374318] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.382896] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.382911] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.391822] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.391836] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.400779] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.400794] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.301 [2024-11-06 08:48:37.409438] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.301 [2024-11-06 08:48:37.409452] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.418515] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.561 [2024-11-06 08:48:37.418530] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.426971] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.561 [2024-11-06 08:48:37.426985] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.436082] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.561 [2024-11-06 08:48:37.436097] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.444665] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.561 [2024-11-06 08:48:37.444680] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.452703] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.561 [2024-11-06 08:48:37.452718] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.461715] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.561 [2024-11-06 08:48:37.461729] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.470797] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.561 [2024-11-06 08:48:37.470812] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.479388] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.561 [2024-11-06 08:48:37.479403] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.487640] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.561 [2024-11-06 08:48:37.487655] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.496688] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.561 [2024-11-06 08:48:37.496703] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.561 [2024-11-06 08:48:37.505211] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.505225] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.513714] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.513729] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.522241] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.522255] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.530532] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.530546] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.539432] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.539446] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.548793] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.548808] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.557235] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.557250] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.566104] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.566119] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.574734] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.574753] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.584066] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.584081] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.592631] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.592645] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.601406] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.601421] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.610445] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.610460] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.619477] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.619492] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.628318] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.628333] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.636477] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.636492] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.645390] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.645405] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.653952] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.653966] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.662501] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.662516] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.562 [2024-11-06 08:48:37.671243] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.562 [2024-11-06 08:48:37.671258] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 18929.00 IOPS, 147.88 MiB/s [2024-11-06T07:48:37.936Z] [2024-11-06 08:48:37.680887] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.680901] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.690086] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.690101] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.697936] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.697951] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.706790] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.706805] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.716033] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.716047] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.723952] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.723966] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.732851] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.732865] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.741579] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.741594] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.750408] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.750422] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.759184] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.759198] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.768084] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.768099] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.777212] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.777227] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.785895] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.785915] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.794536] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.794551] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.803476] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.803491] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.812363] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.812378] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.821434] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.821448] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.830080] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.830095] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.838632] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.838647] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.846789] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.846804] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.856047] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.856062] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.865328] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.865342] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.874449] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.874464] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.883169] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.883185] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.892152] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.892166] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.901231] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.901246] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.910250] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.910265] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.919142] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.919157] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:47.823 [2024-11-06 08:48:37.927833] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:47.823 [2024-11-06 08:48:37.927848] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:37.936866] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:37.936882] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:37.945427] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:37.945442] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:37.954433] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:37.954451] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:37.963337] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:37.963351] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:37.971806] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:37.971820] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:37.980658] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:37.980673] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:37.989343] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:37.989358] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:37.998379] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:37.998394] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.007261] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.007276] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.015916] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.015931] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.024676] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.024690] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.033555] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.033570] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.041528] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.041543] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.050412] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.050427] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.059021] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.059036] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.067664] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.067679] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.076367] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.076382] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.085708] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.085723] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.094521] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.094536] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.103457] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.103472] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.111400] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.111415] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.120918] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.120936] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.129282] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.129297] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.138179] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.138193] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.147116] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.147132] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.156025] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.156040] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.164572] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.164586] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.173233] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.173248] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.205 [2024-11-06 08:48:38.181770] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.205 [2024-11-06 08:48:38.181786] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.190286] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.190300] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.198937] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.198952] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.206869] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.206883] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.215483] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.215498] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.224396] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.224411] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.232928] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.232943] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.241757] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.241772] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.250718] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.250733] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.259617] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.259632] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.268152] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.268167] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.277032] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.277047] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.285812] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.285830] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.294727] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.294742] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.303131] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.303145] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.206 [2024-11-06 08:48:38.311742] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.206 [2024-11-06 08:48:38.311762] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.467 [2024-11-06 08:48:38.320635] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.467 [2024-11-06 08:48:38.320650] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.467 [2024-11-06 08:48:38.329550] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.467 [2024-11-06 08:48:38.329566] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.467 [2024-11-06 08:48:38.338071] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.338085] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.346593] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.346609] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.355510] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.355526] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.364663] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.364678] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.373207] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.373222] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.381662] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.381677] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.390523] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.390537] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.399301] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.399316] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.408580] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.408595] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.417370] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.417385] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.426053] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.426068] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.435190] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.435205] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.443529] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.443543] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.452344] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.452363] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.461038] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.461053] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.469414] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.469429] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.478476] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.478490] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.486797] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.486812] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.495635] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.495649] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.504132] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.504146] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.513506] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.513521] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.521506] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.521521] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.530555] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.530570] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.539337] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.539351] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.548508] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.548523] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.557113] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.557127] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.565896] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.565910] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.468 [2024-11-06 08:48:38.574346] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.468 [2024-11-06 08:48:38.574360] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.583222] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.583236] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.591638] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.591652] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.600483] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.600497] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.608916] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.608930] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.617985] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.617999] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.626742] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.626759] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.635939] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.635954] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.644471] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.644486] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.652874] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.652889] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.661602] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.661616] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.670791] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.670806] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.679684] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.679700] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 18971.00 IOPS, 148.21 MiB/s [2024-11-06T07:48:38.842Z] [2024-11-06 08:48:38.688228] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.688243] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.697303] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.697318] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.706283] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.706298] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.715449] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.715464] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.723411] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.723426] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.731852] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.731867] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.740269] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.740284] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.748806] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.748821] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.757648] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.729 [2024-11-06 08:48:38.757662] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.729 [2024-11-06 08:48:38.766635] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.730 [2024-11-06 08:48:38.766650] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.730 [2024-11-06 08:48:38.775186] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.730 [2024-11-06 08:48:38.775202] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.730 [2024-11-06 08:48:38.783848] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.730 [2024-11-06 08:48:38.783862] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.730 [2024-11-06 08:48:38.793211] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.730 [2024-11-06 08:48:38.793226] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.730 [2024-11-06 08:48:38.801872] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.730 [2024-11-06 08:48:38.801886] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.730 [2024-11-06 08:48:38.809852] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.730 [2024-11-06 08:48:38.809866] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.730 [2024-11-06 08:48:38.818771] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.730 [2024-11-06 08:48:38.818786] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.730 [2024-11-06 08:48:38.827336] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.730 [2024-11-06 08:48:38.827350] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.730 [2024-11-06 08:48:38.835983] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.730 [2024-11-06 08:48:38.835997] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.990 [2024-11-06 08:48:38.844849] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.990 [2024-11-06 08:48:38.844865] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.990 [2024-11-06 08:48:38.853998] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.990 [2024-11-06 08:48:38.854012] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.990 [2024-11-06 08:48:38.862909] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.990 [2024-11-06 08:48:38.862925] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.990 [2024-11-06 08:48:38.871909] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.990 [2024-11-06 08:48:38.871923] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.990 [2024-11-06 08:48:38.880443] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.990 [2024-11-06 08:48:38.880458] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.990 [2024-11-06 08:48:38.889191] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.990 [2024-11-06 08:48:38.889205] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.990 [2024-11-06 08:48:38.897411] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.990 [2024-11-06 08:48:38.897425] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.990 [2024-11-06 08:48:38.906170] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.990 [2024-11-06 08:48:38.906185] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.990 [2024-11-06 08:48:38.915177] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:38.915192] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:38.924285] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:38.924299] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:38.933147] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:38.933162] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:38.941820] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:38.941838] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:38.950566] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:38.950581] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:38.959430] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:38.959444] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:38.967984] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:38.967999] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:38.976907] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:38.976921] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:38.986363] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:38.986378] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:38.995195] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:38.995209] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.003793] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.003808] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.012604] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.012618] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.021370] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.021384] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.030135] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.030150] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.038471] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.038485] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.047144] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.047159] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.055928] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.055942] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.065029] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.065044] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.073983] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.073998] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.083042] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.083056] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.092072] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.092087] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:48.991 [2024-11-06 08:48:39.100778] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:48.991 [2024-11-06 08:48:39.100793] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.251 [2024-11-06 08:48:39.109277] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.251 [2024-11-06 08:48:39.109296] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.251 [2024-11-06 08:48:39.118682] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.251 [2024-11-06 08:48:39.118696] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.251 [2024-11-06 08:48:39.126716] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.251 [2024-11-06 08:48:39.126731] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.251 [2024-11-06 08:48:39.135654] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.251 [2024-11-06 08:48:39.135669] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.251 [2024-11-06 08:48:39.144871] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.144886] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.152972] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.152986] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.161739] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.161757] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.170490] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.170505] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.179703] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.179717] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.188310] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.188325] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.197498] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.197513] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.206037] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.206051] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.214975] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.214989] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.223890] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.223905] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.233086] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.233100] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.241643] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.241658] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.250197] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.250212] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.259059] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.259074] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.267755] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.267770] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.276549] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.276567] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.285778] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.285793] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.294243] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.294257] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.302993] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.303008] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.311864] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.311879] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.320250] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.320265] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.329262] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.329277] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.338337] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.338351] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.346409] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.346424] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.252 [2024-11-06 08:48:39.355278] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.252 [2024-11-06 08:48:39.355292] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.364105] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.364121] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.373059] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.373074] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.382415] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.382430] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.391207] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.391221] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.400296] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.400310] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.408834] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.408848] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.417246] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.417260] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.426488] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.426503] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.434563] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.434577] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.443322] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.443341] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.452365] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.452380] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.460868] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.460882] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.469627] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.469641] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.478571] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.478585] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.487438] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.487453] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.496490] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.496504] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.505618] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.505632] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.514892] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.514907] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.523932] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.523947] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.533062] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.533076] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.541822] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.541837] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.551038] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.551053] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.559688] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.559702] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.568261] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.568275] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.576812] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.576826] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.585620] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.585635] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.594207] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.594222] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.602953] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.602967] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.611792] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.611806] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.513 [2024-11-06 08:48:39.619602] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.513 [2024-11-06 08:48:39.619616] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.773 [2024-11-06 08:48:39.628087] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.773 [2024-11-06 08:48:39.628102] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.773 [2024-11-06 08:48:39.636713] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.773 [2024-11-06 08:48:39.636727] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.645634] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.645649] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.654169] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.654183] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.667718] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.667734] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.675685] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.675699] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 18991.50 IOPS, 148.37 MiB/s [2024-11-06T07:48:39.887Z] [2024-11-06 08:48:39.684830] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.684845] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.693201] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.693215] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.701861] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.701876] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.711043] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.711058] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.719824] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.719839] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.728820] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.728835] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.737410] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.737424] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.746327] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.746341] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.755607] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.755622] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.763622] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.763636] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.772914] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.772928] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.780840] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.780854] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.789631] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.789645] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.798615] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.798629] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.807848] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.807863] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.816236] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.816250] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.825483] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.825498] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.834890] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.834905] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.843775] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.843790] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.852388] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.852403] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.861258] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.861272] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.870358] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.870374] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:49.774 [2024-11-06 08:48:39.878702] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:49.774 [2024-11-06 08:48:39.878717] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.887641] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.887657] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.896132] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.896147] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.904632] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.904646] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.913440] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.913455] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.922234] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.922249] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.931348] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.931362] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.939800] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.939815] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.948230] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.948245] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.957385] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.957400] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.966209] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.966223] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.974633] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.974648] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.983722] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.983738] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:39.992391] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:39.992406] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.000996] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.001012] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.010452] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.010469] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.019493] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.019508] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.028710] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.028725] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.037376] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.037391] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.046225] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.046240] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.055440] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.055454] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.064719] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.064734] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.073348] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.073363] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.081614] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.081629] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.090578] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.090593] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.099479] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.099493] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.107859] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.107878] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.116373] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.116388] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.124999] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.125014] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.133578] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.133593] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.035 [2024-11-06 08:48:40.142259] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.035 [2024-11-06 08:48:40.142274] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.151025] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.151041] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.160022] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.160037] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.169406] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.169421] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.177433] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.177448] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.186266] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.186281] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.195001] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.195016] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.204157] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.204172] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.213261] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.213276] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.221720] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.221735] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.230511] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.230526] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.238756] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.238771] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.248025] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.248040] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.257254] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.257269] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.265628] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.265643] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.274883] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.274901] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.283499] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.283513] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.292366] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.292381] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.301265] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.301279] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.309678] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.309692] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.318536] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.318550] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.327003] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.327018] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.335872] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.335887] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.297 [2024-11-06 08:48:40.344819] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.297 [2024-11-06 08:48:40.344834] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.298 [2024-11-06 08:48:40.353342] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.298 [2024-11-06 08:48:40.353357] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.298 [2024-11-06 08:48:40.361871] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.298 [2024-11-06 08:48:40.361886] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.298 [2024-11-06 08:48:40.370349] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.298 [2024-11-06 08:48:40.370364] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.298 [2024-11-06 08:48:40.378788] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.298 [2024-11-06 08:48:40.378803] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.298 [2024-11-06 08:48:40.387777] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.298 [2024-11-06 08:48:40.387793] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.298 [2024-11-06 08:48:40.396760] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.298 [2024-11-06 08:48:40.396774] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.298 [2024-11-06 08:48:40.405469] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.298 [2024-11-06 08:48:40.405485] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.558 [2024-11-06 08:48:40.414266] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.558 [2024-11-06 08:48:40.414281] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.558 [2024-11-06 08:48:40.422752] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.558 [2024-11-06 08:48:40.422767] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.558 [2024-11-06 08:48:40.432037] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.558 [2024-11-06 08:48:40.432052] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.558 [2024-11-06 08:48:40.440972] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.558 [2024-11-06 08:48:40.440991] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.558 [2024-11-06 08:48:40.449975] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.558 [2024-11-06 08:48:40.449990] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.558 [2024-11-06 08:48:40.458117] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.558 [2024-11-06 08:48:40.458131] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.558 [2024-11-06 08:48:40.467052] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.558 [2024-11-06 08:48:40.467067] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.558 [2024-11-06 08:48:40.476138] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.558 [2024-11-06 08:48:40.476153] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.558 [2024-11-06 08:48:40.484458] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.484473] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.493191] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.493206] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.502387] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.502402] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.511386] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.511400] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.520458] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.520473] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.528899] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.528915] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.537774] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.537788] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.546606] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.546621] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.555026] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.555041] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.564311] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.564326] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.572384] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.572399] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.581558] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.581573] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.590421] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.590436] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.599470] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.599484] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.608473] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.608492] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.616908] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.616923] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.625476] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.625490] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.634637] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.634652] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.643268] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.643282] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.652058] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.652073] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.660777] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.660791] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.559 [2024-11-06 08:48:40.669791] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.559 [2024-11-06 08:48:40.669806] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.678896] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.678911] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 18993.60 IOPS, 148.39 MiB/s [2024-11-06T07:48:40.933Z] [2024-11-06 08:48:40.686115] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.686130] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 00:12:50.820 Latency(us) 00:12:50.820 [2024-11-06T07:48:40.933Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:50.820 Job: Nvme1n1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 128, IO size: 8192) 00:12:50.820 Nvme1n1 : 5.00 19000.64 148.44 0.00 0.00 6730.20 2621.44 17367.04 00:12:50.820 [2024-11-06T07:48:40.933Z] =================================================================================================================== 00:12:50.820 [2024-11-06T07:48:40.933Z] Total : 19000.64 148.44 0.00 0.00 6730.20 2621.44 17367.04 00:12:50.820 [2024-11-06 08:48:40.693201] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.693214] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.701218] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.701229] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.709239] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.709249] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.717261] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.717272] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.725279] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.725288] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.733297] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.733305] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.741319] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.741328] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.749340] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.749348] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.757358] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.757366] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.765378] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.765385] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.773398] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.773405] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.781420] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.781429] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.789439] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.789447] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 [2024-11-06 08:48:40.797460] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:12:50.820 [2024-11-06 08:48:40.797467] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:12:50.820 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh: line 42: kill: (1739687) - No such process 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@49 -- # wait 1739687 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@52 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@53 -- # rpc_cmd bdev_delay_create -b malloc0 -d delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:50.820 delay0 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@54 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 delay0 -n 1 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:12:50.820 08:48:40 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -c 0x1 -t 5 -q 64 -w randrw -M 50 -l warning -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 ns:1' 00:12:51.080 [2024-11-06 08:48:40.984955] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:12:57.662 [2024-11-06 08:48:47.105420] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x941560 is same with the state(6) to be set 00:12:57.662 [2024-11-06 08:48:47.105455] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x941560 is same with the state(6) to be set 00:12:57.662 Initializing NVMe Controllers 00:12:57.662 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:12:57.662 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:12:57.662 Initialization complete. Launching workers. 00:12:57.662 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 I/O completed: 320, failed: 669 00:12:57.662 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) abort submitted 959, failed to submit 30 00:12:57.662 success 728, unsuccessful 231, failed 0 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@59 -- # trap - SIGINT SIGTERM EXIT 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- target/zcopy.sh@60 -- # nvmftestfini 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@514 -- # nvmfcleanup 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@121 -- # sync 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@124 -- # set +e 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@125 -- # for i in {1..20} 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:12:57.663 rmmod nvme_tcp 00:12:57.663 rmmod nvme_fabrics 00:12:57.663 rmmod nvme_keyring 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@128 -- # set -e 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@129 -- # return 0 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@515 -- # '[' -n 1737412 ']' 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@516 -- # killprocess 1737412 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@950 -- # '[' -z 1737412 ']' 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@954 -- # kill -0 1737412 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@955 -- # uname 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1737412 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1737412' 00:12:57.663 killing process with pid 1737412 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@969 -- # kill 1737412 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@974 -- # wait 1737412 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@297 -- # iptr 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@789 -- # iptables-save 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@789 -- # iptables-restore 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@302 -- # remove_spdk_ns 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:57.663 08:48:47 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:12:59.574 00:12:59.574 real 0m33.257s 00:12:59.574 user 0m44.785s 00:12:59.574 sys 0m10.784s 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@1126 -- # xtrace_disable 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:12:59.574 ************************************ 00:12:59.574 END TEST nvmf_zcopy 00:12:59.574 ************************************ 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@33 -- # run_test nvmf_nmic /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:12:59.574 ************************************ 00:12:59.574 START TEST nvmf_nmic 00:12:59.574 ************************************ 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:12:59.574 * Looking for test storage... 00:12:59.574 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:12:59.574 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1689 -- # lcov --version 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@333 -- # local ver1 ver1_l 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@334 -- # local ver2 ver2_l 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@336 -- # IFS=.-: 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@336 -- # read -ra ver1 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@337 -- # IFS=.-: 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@337 -- # read -ra ver2 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@338 -- # local 'op=<' 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@340 -- # ver1_l=2 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@341 -- # ver2_l=1 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@344 -- # case "$op" in 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@345 -- # : 1 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@364 -- # (( v = 0 )) 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@365 -- # decimal 1 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@353 -- # local d=1 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@355 -- # echo 1 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@365 -- # ver1[v]=1 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@366 -- # decimal 2 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@353 -- # local d=2 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@355 -- # echo 2 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@366 -- # ver2[v]=2 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@368 -- # return 0 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:12:59.835 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:59.835 --rc genhtml_branch_coverage=1 00:12:59.835 --rc genhtml_function_coverage=1 00:12:59.835 --rc genhtml_legend=1 00:12:59.835 --rc geninfo_all_blocks=1 00:12:59.835 --rc geninfo_unexecuted_blocks=1 00:12:59.835 00:12:59.835 ' 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:12:59.835 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:59.835 --rc genhtml_branch_coverage=1 00:12:59.835 --rc genhtml_function_coverage=1 00:12:59.835 --rc genhtml_legend=1 00:12:59.835 --rc geninfo_all_blocks=1 00:12:59.835 --rc geninfo_unexecuted_blocks=1 00:12:59.835 00:12:59.835 ' 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:12:59.835 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:59.835 --rc genhtml_branch_coverage=1 00:12:59.835 --rc genhtml_function_coverage=1 00:12:59.835 --rc genhtml_legend=1 00:12:59.835 --rc geninfo_all_blocks=1 00:12:59.835 --rc geninfo_unexecuted_blocks=1 00:12:59.835 00:12:59.835 ' 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:12:59.835 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:12:59.835 --rc genhtml_branch_coverage=1 00:12:59.835 --rc genhtml_function_coverage=1 00:12:59.835 --rc genhtml_legend=1 00:12:59.835 --rc geninfo_all_blocks=1 00:12:59.835 --rc geninfo_unexecuted_blocks=1 00:12:59.835 00:12:59.835 ' 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@7 -- # uname -s 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@15 -- # shopt -s extglob 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@5 -- # export PATH 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@51 -- # : 0 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:12:59.835 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@55 -- # have_pci_nics=0 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@11 -- # MALLOC_BDEV_SIZE=64 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:12:59.835 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@14 -- # nvmftestinit 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@474 -- # prepare_net_devs 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@436 -- # local -g is_hw=no 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@438 -- # remove_spdk_ns 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@309 -- # xtrace_disable 00:12:59.836 08:48:49 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@315 -- # pci_devs=() 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@315 -- # local -a pci_devs 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@316 -- # pci_net_devs=() 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@317 -- # pci_drivers=() 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@317 -- # local -A pci_drivers 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@319 -- # net_devs=() 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@319 -- # local -ga net_devs 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@320 -- # e810=() 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@320 -- # local -ga e810 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@321 -- # x722=() 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@321 -- # local -ga x722 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@322 -- # mlx=() 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@322 -- # local -ga mlx 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:13:07.974 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:13:07.974 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:13:07.974 Found net devices under 0000:4b:00.0: cvl_0_0 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:07.974 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:13:07.975 Found net devices under 0000:4b:00.1: cvl_0_1 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@440 -- # is_hw=yes 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:13:07.975 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:07.975 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.626 ms 00:13:07.975 00:13:07.975 --- 10.0.0.2 ping statistics --- 00:13:07.975 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:07.975 rtt min/avg/max/mdev = 0.626/0.626/0.626/0.000 ms 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:07.975 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:07.975 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.304 ms 00:13:07.975 00:13:07.975 --- 10.0.0.1 ping statistics --- 00:13:07.975 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:07.975 rtt min/avg/max/mdev = 0.304/0.304/0.304/0.000 ms 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@448 -- # return 0 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:13:07.975 08:48:56 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@15 -- # nvmfappstart -m 0xF 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@724 -- # xtrace_disable 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@507 -- # nvmfpid=1746144 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@508 -- # waitforlisten 1746144 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@831 -- # '[' -z 1746144 ']' 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:07.975 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.975 [2024-11-06 08:48:57.083527] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:13:07.975 [2024-11-06 08:48:57.083592] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:07.975 [2024-11-06 08:48:57.166782] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:07.975 [2024-11-06 08:48:57.210170] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:07.975 [2024-11-06 08:48:57.210208] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:07.975 [2024-11-06 08:48:57.210217] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:07.975 [2024-11-06 08:48:57.210224] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:07.975 [2024-11-06 08:48:57.210229] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:07.975 [2024-11-06 08:48:57.211788] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:13:07.975 [2024-11-06 08:48:57.212013] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:07.975 [2024-11-06 08:48:57.212013] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:13:07.975 [2024-11-06 08:48:57.211868] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@864 -- # return 0 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@730 -- # xtrace_disable 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.975 [2024-11-06 08:48:57.940358] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.975 Malloc0 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@21 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.975 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.976 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.976 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:07.976 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.976 08:48:57 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.976 [2024-11-06 08:48:58.007062] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@25 -- # echo 'test case1: single bdev can'\''t be used in multiple subsystems' 00:13:07.976 test case1: single bdev can't be used in multiple subsystems 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@26 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@28 -- # nmic_status=0 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc0 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.976 [2024-11-06 08:48:58.043011] bdev.c:8456:bdev_open: *ERROR*: bdev Malloc0 already claimed: type exclusive_write by module NVMe-oF Target 00:13:07.976 [2024-11-06 08:48:58.043031] subsystem.c:2151:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2: bdev Malloc0 cannot be opened, error=-1 00:13:07.976 [2024-11-06 08:48:58.043039] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:13:07.976 request: 00:13:07.976 { 00:13:07.976 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:13:07.976 "namespace": { 00:13:07.976 "bdev_name": "Malloc0", 00:13:07.976 "no_auto_visible": false 00:13:07.976 }, 00:13:07.976 "method": "nvmf_subsystem_add_ns", 00:13:07.976 "req_id": 1 00:13:07.976 } 00:13:07.976 Got JSON-RPC error response 00:13:07.976 response: 00:13:07.976 { 00:13:07.976 "code": -32602, 00:13:07.976 "message": "Invalid parameters" 00:13:07.976 } 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@29 -- # nmic_status=1 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@31 -- # '[' 1 -eq 0 ']' 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@36 -- # echo ' Adding namespace failed - expected result.' 00:13:07.976 Adding namespace failed - expected result. 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@39 -- # echo 'test case2: host connect to nvmf target in multiple paths' 00:13:07.976 test case2: host connect to nvmf target in multiple paths 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:07.976 [2024-11-06 08:48:58.055146] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:07.976 08:48:58 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@41 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:09.886 08:48:59 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@42 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4421 00:13:11.269 08:49:01 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@44 -- # waitforserial SPDKISFASTANDAWESOME 00:13:11.269 08:49:01 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1198 -- # local i=0 00:13:11.269 08:49:01 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:13:11.269 08:49:01 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:13:11.269 08:49:01 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1205 -- # sleep 2 00:13:13.180 08:49:03 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:13:13.180 08:49:03 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:13:13.180 08:49:03 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:13:13.180 08:49:03 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:13:13.180 08:49:03 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:13:13.180 08:49:03 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1208 -- # return 0 00:13:13.180 08:49:03 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:13:13.180 [global] 00:13:13.180 thread=1 00:13:13.180 invalidate=1 00:13:13.180 rw=write 00:13:13.180 time_based=1 00:13:13.180 runtime=1 00:13:13.180 ioengine=libaio 00:13:13.180 direct=1 00:13:13.180 bs=4096 00:13:13.180 iodepth=1 00:13:13.180 norandommap=0 00:13:13.180 numjobs=1 00:13:13.180 00:13:13.180 verify_dump=1 00:13:13.180 verify_backlog=512 00:13:13.180 verify_state_save=0 00:13:13.180 do_verify=1 00:13:13.180 verify=crc32c-intel 00:13:13.180 [job0] 00:13:13.180 filename=/dev/nvme0n1 00:13:13.180 Could not set queue depth (nvme0n1) 00:13:13.749 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:13.749 fio-3.35 00:13:13.749 Starting 1 thread 00:13:14.689 00:13:14.689 job0: (groupid=0, jobs=1): err= 0: pid=1747696: Wed Nov 6 08:49:04 2024 00:13:14.689 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:13:14.689 slat (nsec): min=7187, max=60834, avg=26269.98, stdev=3451.96 00:13:14.689 clat (usec): min=574, max=1150, avg=966.66, stdev=75.09 00:13:14.689 lat (usec): min=600, max=1176, avg=992.93, stdev=75.23 00:13:14.689 clat percentiles (usec): 00:13:14.689 | 1.00th=[ 750], 5.00th=[ 799], 10.00th=[ 857], 20.00th=[ 922], 00:13:14.689 | 30.00th=[ 955], 40.00th=[ 971], 50.00th=[ 979], 60.00th=[ 996], 00:13:14.689 | 70.00th=[ 1004], 80.00th=[ 1020], 90.00th=[ 1045], 95.00th=[ 1074], 00:13:14.689 | 99.00th=[ 1106], 99.50th=[ 1123], 99.90th=[ 1156], 99.95th=[ 1156], 00:13:14.689 | 99.99th=[ 1156] 00:13:14.689 write: IOPS=731, BW=2925KiB/s (2995kB/s)(2928KiB/1001msec); 0 zone resets 00:13:14.689 slat (usec): min=9, max=30642, avg=71.08, stdev=1131.55 00:13:14.689 clat (usec): min=259, max=804, avg=588.31, stdev=92.99 00:13:14.689 lat (usec): min=269, max=31288, avg=659.39, stdev=1137.90 00:13:14.689 clat percentiles (usec): 00:13:14.689 | 1.00th=[ 355], 5.00th=[ 412], 10.00th=[ 453], 20.00th=[ 502], 00:13:14.689 | 30.00th=[ 553], 40.00th=[ 578], 50.00th=[ 594], 60.00th=[ 619], 00:13:14.689 | 70.00th=[ 644], 80.00th=[ 676], 90.00th=[ 701], 95.00th=[ 725], 00:13:14.689 | 99.00th=[ 766], 99.50th=[ 775], 99.90th=[ 807], 99.95th=[ 807], 00:13:14.689 | 99.99th=[ 807] 00:13:14.690 bw ( KiB/s): min= 4096, max= 4096, per=100.00%, avg=4096.00, stdev= 0.00, samples=1 00:13:14.690 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:14.690 lat (usec) : 500=11.01%, 750=47.19%, 1000=28.14% 00:13:14.690 lat (msec) : 2=13.67% 00:13:14.690 cpu : usr=1.90%, sys=3.50%, ctx=1248, majf=0, minf=1 00:13:14.690 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:14.690 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:14.690 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:14.690 issued rwts: total=512,732,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:14.690 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:14.690 00:13:14.690 Run status group 0 (all jobs): 00:13:14.690 READ: bw=2046KiB/s (2095kB/s), 2046KiB/s-2046KiB/s (2095kB/s-2095kB/s), io=2048KiB (2097kB), run=1001-1001msec 00:13:14.690 WRITE: bw=2925KiB/s (2995kB/s), 2925KiB/s-2925KiB/s (2995kB/s-2995kB/s), io=2928KiB (2998kB), run=1001-1001msec 00:13:14.690 00:13:14.690 Disk stats (read/write): 00:13:14.690 nvme0n1: ios=537/568, merge=0/0, ticks=1460/315, in_queue=1775, util=98.90% 00:13:14.690 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@48 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:14.950 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 2 controller(s) 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@49 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1219 -- # local i=0 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1231 -- # return 0 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- target/nmic.sh@53 -- # nvmftestfini 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@514 -- # nvmfcleanup 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@121 -- # sync 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@124 -- # set +e 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:14.950 08:49:04 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:14.950 rmmod nvme_tcp 00:13:14.950 rmmod nvme_fabrics 00:13:14.950 rmmod nvme_keyring 00:13:14.950 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:14.950 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@128 -- # set -e 00:13:14.950 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@129 -- # return 0 00:13:14.950 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@515 -- # '[' -n 1746144 ']' 00:13:14.950 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@516 -- # killprocess 1746144 00:13:14.950 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@950 -- # '[' -z 1746144 ']' 00:13:14.950 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@954 -- # kill -0 1746144 00:13:14.950 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@955 -- # uname 00:13:14.950 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:14.950 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1746144 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1746144' 00:13:15.211 killing process with pid 1746144 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@969 -- # kill 1746144 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@974 -- # wait 1746144 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@297 -- # iptr 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@789 -- # iptables-save 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@789 -- # iptables-restore 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:15.211 08:49:05 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_nmic -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:17.754 00:13:17.754 real 0m17.750s 00:13:17.754 user 0m45.697s 00:13:17.754 sys 0m6.338s 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:13:17.754 ************************************ 00:13:17.754 END TEST nvmf_nmic 00:13:17.754 ************************************ 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@34 -- # run_test nvmf_fio_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:13:17.754 ************************************ 00:13:17.754 START TEST nvmf_fio_target 00:13:17.754 ************************************ 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:13:17.754 * Looking for test storage... 00:13:17.754 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1689 -- # lcov --version 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@336 -- # IFS=.-: 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@336 -- # read -ra ver1 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@337 -- # IFS=.-: 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@337 -- # read -ra ver2 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@338 -- # local 'op=<' 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@340 -- # ver1_l=2 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@341 -- # ver2_l=1 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@344 -- # case "$op" in 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@345 -- # : 1 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@365 -- # decimal 1 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@353 -- # local d=1 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@355 -- # echo 1 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@365 -- # ver1[v]=1 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@366 -- # decimal 2 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@353 -- # local d=2 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@355 -- # echo 2 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@366 -- # ver2[v]=2 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@368 -- # return 0 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:13:17.754 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:17.754 --rc genhtml_branch_coverage=1 00:13:17.754 --rc genhtml_function_coverage=1 00:13:17.754 --rc genhtml_legend=1 00:13:17.754 --rc geninfo_all_blocks=1 00:13:17.754 --rc geninfo_unexecuted_blocks=1 00:13:17.754 00:13:17.754 ' 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:13:17.754 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:17.754 --rc genhtml_branch_coverage=1 00:13:17.754 --rc genhtml_function_coverage=1 00:13:17.754 --rc genhtml_legend=1 00:13:17.754 --rc geninfo_all_blocks=1 00:13:17.754 --rc geninfo_unexecuted_blocks=1 00:13:17.754 00:13:17.754 ' 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:13:17.754 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:17.754 --rc genhtml_branch_coverage=1 00:13:17.754 --rc genhtml_function_coverage=1 00:13:17.754 --rc genhtml_legend=1 00:13:17.754 --rc geninfo_all_blocks=1 00:13:17.754 --rc geninfo_unexecuted_blocks=1 00:13:17.754 00:13:17.754 ' 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:13:17.754 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:17.754 --rc genhtml_branch_coverage=1 00:13:17.754 --rc genhtml_function_coverage=1 00:13:17.754 --rc genhtml_legend=1 00:13:17.754 --rc geninfo_all_blocks=1 00:13:17.754 --rc geninfo_unexecuted_blocks=1 00:13:17.754 00:13:17.754 ' 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@7 -- # uname -s 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:17.754 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@15 -- # shopt -s extglob 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@5 -- # export PATH 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@51 -- # : 0 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:17.755 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@16 -- # nvmftestinit 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@474 -- # prepare_net_devs 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@436 -- # local -g is_hw=no 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@438 -- # remove_spdk_ns 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@309 -- # xtrace_disable 00:13:17.755 08:49:07 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@315 -- # pci_devs=() 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@319 -- # net_devs=() 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@320 -- # e810=() 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@320 -- # local -ga e810 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@321 -- # x722=() 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@321 -- # local -ga x722 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@322 -- # mlx=() 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@322 -- # local -ga mlx 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:13:25.895 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:13:25.895 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:13:25.895 Found net devices under 0000:4b:00.0: cvl_0_0 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:13:25.895 Found net devices under 0000:4b:00.1: cvl_0_1 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@440 -- # is_hw=yes 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:25.895 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:13:25.896 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:25.896 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.648 ms 00:13:25.896 00:13:25.896 --- 10.0.0.2 ping statistics --- 00:13:25.896 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:25.896 rtt min/avg/max/mdev = 0.648/0.648/0.648/0.000 ms 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:25.896 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:25.896 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.299 ms 00:13:25.896 00:13:25.896 --- 10.0.0.1 ping statistics --- 00:13:25.896 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:25.896 rtt min/avg/max/mdev = 0.299/0.299/0.299/0.000 ms 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@448 -- # return 0 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@17 -- # nvmfappstart -m 0xF 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@724 -- # xtrace_disable 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@507 -- # nvmfpid=1752164 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@508 -- # waitforlisten 1752164 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@831 -- # '[' -z 1752164 ']' 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:25.896 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:25.896 08:49:14 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:13:25.896 [2024-11-06 08:49:14.873275] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:13:25.896 [2024-11-06 08:49:14.873345] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:25.896 [2024-11-06 08:49:14.961325] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:25.896 [2024-11-06 08:49:15.002842] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:25.896 [2024-11-06 08:49:15.002880] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:25.896 [2024-11-06 08:49:15.002888] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:25.896 [2024-11-06 08:49:15.002895] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:25.896 [2024-11-06 08:49:15.002901] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:25.896 [2024-11-06 08:49:15.004499] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:13:25.896 [2024-11-06 08:49:15.004622] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:13:25.896 [2024-11-06 08:49:15.004797] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:13:25.896 [2024-11-06 08:49:15.004798] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:25.896 08:49:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:25.896 08:49:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@864 -- # return 0 00:13:25.896 08:49:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:13:25.896 08:49:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@730 -- # xtrace_disable 00:13:25.896 08:49:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:13:25.896 08:49:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:25.896 08:49:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:13:25.896 [2024-11-06 08:49:15.877524] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:25.896 08:49:15 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:26.157 08:49:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@21 -- # malloc_bdevs='Malloc0 ' 00:13:26.157 08:49:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:26.418 08:49:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@22 -- # malloc_bdevs+=Malloc1 00:13:26.418 08:49:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:26.418 08:49:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@24 -- # raid_malloc_bdevs='Malloc2 ' 00:13:26.418 08:49:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:26.679 08:49:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@25 -- # raid_malloc_bdevs+=Malloc3 00:13:26.679 08:49:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc2 Malloc3' 00:13:26.939 08:49:16 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:27.197 08:49:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@29 -- # concat_malloc_bdevs='Malloc4 ' 00:13:27.197 08:49:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:27.197 08:49:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@30 -- # concat_malloc_bdevs+='Malloc5 ' 00:13:27.197 08:49:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:13:27.457 08:49:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@31 -- # concat_malloc_bdevs+=Malloc6 00:13:27.457 08:49:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n concat0 -r concat -z 64 -b 'Malloc4 Malloc5 Malloc6' 00:13:27.718 08:49:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:13:27.718 08:49:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:13:27.718 08:49:17 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:13:27.979 08:49:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:13:27.979 08:49:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:13:28.238 08:49:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:28.238 [2024-11-06 08:49:18.344030] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:28.498 08:49:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 raid0 00:13:28.498 08:49:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 concat0 00:13:28.758 08:49:18 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@46 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:13:30.137 08:49:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@48 -- # waitforserial SPDKISFASTANDAWESOME 4 00:13:30.137 08:49:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1198 -- # local i=0 00:13:30.137 08:49:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:13:30.137 08:49:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1200 -- # [[ -n 4 ]] 00:13:30.137 08:49:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1201 -- # nvme_device_counter=4 00:13:30.137 08:49:20 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1205 -- # sleep 2 00:13:32.677 08:49:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:13:32.677 08:49:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:13:32.677 08:49:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:13:32.677 08:49:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1207 -- # nvme_devices=4 00:13:32.677 08:49:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:13:32.677 08:49:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1208 -- # return 0 00:13:32.677 08:49:22 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:13:32.677 [global] 00:13:32.677 thread=1 00:13:32.677 invalidate=1 00:13:32.677 rw=write 00:13:32.677 time_based=1 00:13:32.677 runtime=1 00:13:32.677 ioengine=libaio 00:13:32.677 direct=1 00:13:32.677 bs=4096 00:13:32.677 iodepth=1 00:13:32.677 norandommap=0 00:13:32.677 numjobs=1 00:13:32.677 00:13:32.677 verify_dump=1 00:13:32.677 verify_backlog=512 00:13:32.677 verify_state_save=0 00:13:32.677 do_verify=1 00:13:32.677 verify=crc32c-intel 00:13:32.677 [job0] 00:13:32.677 filename=/dev/nvme0n1 00:13:32.677 [job1] 00:13:32.677 filename=/dev/nvme0n2 00:13:32.677 [job2] 00:13:32.677 filename=/dev/nvme0n3 00:13:32.677 [job3] 00:13:32.677 filename=/dev/nvme0n4 00:13:32.677 Could not set queue depth (nvme0n1) 00:13:32.677 Could not set queue depth (nvme0n2) 00:13:32.677 Could not set queue depth (nvme0n3) 00:13:32.677 Could not set queue depth (nvme0n4) 00:13:32.677 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:32.677 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:32.677 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:32.677 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:32.677 fio-3.35 00:13:32.677 Starting 4 threads 00:13:34.055 00:13:34.055 job0: (groupid=0, jobs=1): err= 0: pid=1753958: Wed Nov 6 08:49:23 2024 00:13:34.055 read: IOPS=46, BW=184KiB/s (189kB/s)(192KiB/1042msec) 00:13:34.055 slat (nsec): min=7004, max=28890, avg=21669.33, stdev=9164.09 00:13:34.055 clat (usec): min=522, max=42936, avg=14485.69, stdev=19652.62 00:13:34.055 lat (usec): min=531, max=42963, avg=14507.36, stdev=19656.14 00:13:34.055 clat percentiles (usec): 00:13:34.055 | 1.00th=[ 523], 5.00th=[ 578], 10.00th=[ 594], 20.00th=[ 652], 00:13:34.055 | 30.00th=[ 734], 40.00th=[ 783], 50.00th=[ 816], 60.00th=[ 873], 00:13:34.055 | 70.00th=[41157], 80.00th=[41681], 90.00th=[42206], 95.00th=[42730], 00:13:34.055 | 99.00th=[42730], 99.50th=[42730], 99.90th=[42730], 99.95th=[42730], 00:13:34.055 | 99.99th=[42730] 00:13:34.055 write: IOPS=491, BW=1965KiB/s (2013kB/s)(2048KiB/1042msec); 0 zone resets 00:13:34.055 slat (nsec): min=9616, max=56359, avg=33655.54, stdev=8585.00 00:13:34.055 clat (usec): min=211, max=960, avg=626.93, stdev=140.02 00:13:34.055 lat (usec): min=224, max=996, avg=660.59, stdev=142.41 00:13:34.055 clat percentiles (usec): 00:13:34.055 | 1.00th=[ 310], 5.00th=[ 400], 10.00th=[ 449], 20.00th=[ 502], 00:13:34.055 | 30.00th=[ 553], 40.00th=[ 586], 50.00th=[ 627], 60.00th=[ 660], 00:13:34.055 | 70.00th=[ 701], 80.00th=[ 750], 90.00th=[ 799], 95.00th=[ 865], 00:13:34.055 | 99.00th=[ 938], 99.50th=[ 955], 99.90th=[ 963], 99.95th=[ 963], 00:13:34.055 | 99.99th=[ 963] 00:13:34.055 bw ( KiB/s): min= 4096, max= 4096, per=37.54%, avg=4096.00, stdev= 0.00, samples=1 00:13:34.055 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:34.055 lat (usec) : 250=0.36%, 500=17.68%, 750=58.57%, 1000=20.54% 00:13:34.055 lat (msec) : 50=2.86% 00:13:34.055 cpu : usr=1.15%, sys=2.21%, ctx=561, majf=0, minf=1 00:13:34.055 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:34.055 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:34.055 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:34.055 issued rwts: total=48,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:34.055 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:34.055 job1: (groupid=0, jobs=1): err= 0: pid=1753959: Wed Nov 6 08:49:23 2024 00:13:34.055 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:13:34.055 slat (nsec): min=7538, max=73666, avg=27884.51, stdev=4481.32 00:13:34.055 clat (usec): min=325, max=1326, avg=973.88, stdev=154.18 00:13:34.055 lat (usec): min=353, max=1353, avg=1001.76, stdev=154.56 00:13:34.055 clat percentiles (usec): 00:13:34.055 | 1.00th=[ 603], 5.00th=[ 693], 10.00th=[ 758], 20.00th=[ 840], 00:13:34.055 | 30.00th=[ 906], 40.00th=[ 947], 50.00th=[ 996], 60.00th=[ 1037], 00:13:34.055 | 70.00th=[ 1074], 80.00th=[ 1106], 90.00th=[ 1156], 95.00th=[ 1205], 00:13:34.055 | 99.00th=[ 1287], 99.50th=[ 1303], 99.90th=[ 1319], 99.95th=[ 1319], 00:13:34.055 | 99.99th=[ 1319] 00:13:34.055 write: IOPS=793, BW=3173KiB/s (3249kB/s)(3176KiB/1001msec); 0 zone resets 00:13:34.055 slat (nsec): min=9305, max=70646, avg=33585.67, stdev=8856.93 00:13:34.055 clat (usec): min=214, max=879, avg=562.99, stdev=133.88 00:13:34.055 lat (usec): min=224, max=915, avg=596.58, stdev=136.79 00:13:34.055 clat percentiles (usec): 00:13:34.055 | 1.00th=[ 239], 5.00th=[ 347], 10.00th=[ 388], 20.00th=[ 449], 00:13:34.055 | 30.00th=[ 486], 40.00th=[ 529], 50.00th=[ 570], 60.00th=[ 603], 00:13:34.055 | 70.00th=[ 635], 80.00th=[ 676], 90.00th=[ 742], 95.00th=[ 783], 00:13:34.055 | 99.00th=[ 848], 99.50th=[ 857], 99.90th=[ 881], 99.95th=[ 881], 00:13:34.055 | 99.99th=[ 881] 00:13:34.055 bw ( KiB/s): min= 4096, max= 4096, per=37.54%, avg=4096.00, stdev= 0.00, samples=1 00:13:34.055 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:34.055 lat (usec) : 250=0.77%, 500=19.91%, 750=38.21%, 1000=21.90% 00:13:34.055 lat (msec) : 2=19.22% 00:13:34.055 cpu : usr=3.30%, sys=4.90%, ctx=1309, majf=0, minf=1 00:13:34.055 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:34.055 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:34.055 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:34.055 issued rwts: total=512,794,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:34.055 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:34.055 job2: (groupid=0, jobs=1): err= 0: pid=1753960: Wed Nov 6 08:49:23 2024 00:13:34.055 read: IOPS=221, BW=887KiB/s (908kB/s)(920KiB/1037msec) 00:13:34.055 slat (nsec): min=6227, max=45697, avg=23353.72, stdev=7608.23 00:13:34.055 clat (usec): min=383, max=42509, avg=3722.00, stdev=10381.08 00:13:34.055 lat (usec): min=410, max=42536, avg=3745.35, stdev=10382.06 00:13:34.055 clat percentiles (usec): 00:13:34.055 | 1.00th=[ 490], 5.00th=[ 586], 10.00th=[ 660], 20.00th=[ 725], 00:13:34.055 | 30.00th=[ 783], 40.00th=[ 807], 50.00th=[ 848], 60.00th=[ 906], 00:13:34.055 | 70.00th=[ 922], 80.00th=[ 947], 90.00th=[ 1012], 95.00th=[41157], 00:13:34.055 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42730], 99.95th=[42730], 00:13:34.055 | 99.99th=[42730] 00:13:34.055 write: IOPS=493, BW=1975KiB/s (2022kB/s)(2048KiB/1037msec); 0 zone resets 00:13:34.055 slat (nsec): min=2924, max=38965, avg=12216.60, stdev=6764.56 00:13:34.055 clat (usec): min=118, max=682, avg=317.87, stdev=113.51 00:13:34.055 lat (usec): min=128, max=712, avg=330.08, stdev=115.26 00:13:34.055 clat percentiles (usec): 00:13:34.055 | 1.00th=[ 126], 5.00th=[ 133], 10.00th=[ 145], 20.00th=[ 217], 00:13:34.056 | 30.00th=[ 251], 40.00th=[ 289], 50.00th=[ 322], 60.00th=[ 351], 00:13:34.056 | 70.00th=[ 379], 80.00th=[ 408], 90.00th=[ 453], 95.00th=[ 515], 00:13:34.056 | 99.00th=[ 603], 99.50th=[ 660], 99.90th=[ 685], 99.95th=[ 685], 00:13:34.056 | 99.99th=[ 685] 00:13:34.056 bw ( KiB/s): min= 4096, max= 4096, per=37.54%, avg=4096.00, stdev= 0.00, samples=1 00:13:34.056 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:34.056 lat (usec) : 250=20.49%, 500=44.61%, 750=11.46%, 1000=20.08% 00:13:34.056 lat (msec) : 2=1.08%, 20=0.13%, 50=2.16% 00:13:34.056 cpu : usr=0.58%, sys=1.25%, ctx=743, majf=0, minf=1 00:13:34.056 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:34.056 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:34.056 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:34.056 issued rwts: total=230,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:34.056 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:34.056 job3: (groupid=0, jobs=1): err= 0: pid=1753961: Wed Nov 6 08:49:23 2024 00:13:34.056 read: IOPS=986, BW=3944KiB/s (4039kB/s)(3948KiB/1001msec) 00:13:34.056 slat (nsec): min=7239, max=63500, avg=24955.86, stdev=7607.12 00:13:34.056 clat (usec): min=152, max=701, avg=495.33, stdev=82.01 00:13:34.056 lat (usec): min=160, max=729, avg=520.29, stdev=81.94 00:13:34.056 clat percentiles (usec): 00:13:34.056 | 1.00th=[ 221], 5.00th=[ 355], 10.00th=[ 392], 20.00th=[ 441], 00:13:34.056 | 30.00th=[ 465], 40.00th=[ 482], 50.00th=[ 498], 60.00th=[ 519], 00:13:34.056 | 70.00th=[ 537], 80.00th=[ 562], 90.00th=[ 594], 95.00th=[ 619], 00:13:34.056 | 99.00th=[ 668], 99.50th=[ 693], 99.90th=[ 701], 99.95th=[ 701], 00:13:34.056 | 99.99th=[ 701] 00:13:34.056 write: IOPS=1022, BW=4092KiB/s (4190kB/s)(4096KiB/1001msec); 0 zone resets 00:13:34.056 slat (usec): min=5, max=31238, avg=55.61, stdev=976.53 00:13:34.056 clat (usec): min=98, max=1347, avg=406.17, stdev=150.35 00:13:34.056 lat (usec): min=109, max=31932, avg=461.77, stdev=997.25 00:13:34.056 clat percentiles (usec): 00:13:34.056 | 1.00th=[ 109], 5.00th=[ 155], 10.00th=[ 239], 20.00th=[ 277], 00:13:34.056 | 30.00th=[ 314], 40.00th=[ 363], 50.00th=[ 392], 60.00th=[ 429], 00:13:34.056 | 70.00th=[ 490], 80.00th=[ 537], 90.00th=[ 611], 95.00th=[ 668], 00:13:34.056 | 99.00th=[ 742], 99.50th=[ 750], 99.90th=[ 799], 99.95th=[ 1352], 00:13:34.056 | 99.99th=[ 1352] 00:13:34.056 bw ( KiB/s): min= 4096, max= 4096, per=37.54%, avg=4096.00, stdev= 0.00, samples=1 00:13:34.056 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:34.056 lat (usec) : 100=0.10%, 250=7.41%, 500=54.15%, 750=37.94%, 1000=0.35% 00:13:34.056 lat (msec) : 2=0.05% 00:13:34.056 cpu : usr=3.30%, sys=4.80%, ctx=2018, majf=0, minf=1 00:13:34.056 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:34.056 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:34.056 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:34.056 issued rwts: total=987,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:34.056 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:34.056 00:13:34.056 Run status group 0 (all jobs): 00:13:34.056 READ: bw=6821KiB/s (6985kB/s), 184KiB/s-3944KiB/s (189kB/s-4039kB/s), io=7108KiB (7279kB), run=1001-1042msec 00:13:34.056 WRITE: bw=10.7MiB/s (11.2MB/s), 1965KiB/s-4092KiB/s (2013kB/s-4190kB/s), io=11.1MiB (11.6MB), run=1001-1042msec 00:13:34.056 00:13:34.056 Disk stats (read/write): 00:13:34.056 nvme0n1: ios=99/512, merge=0/0, ticks=560/252, in_queue=812, util=87.07% 00:13:34.056 nvme0n2: ios=556/512, merge=0/0, ticks=1062/236, in_queue=1298, util=87.54% 00:13:34.056 nvme0n3: ios=282/512, merge=0/0, ticks=1235/160, in_queue=1395, util=91.54% 00:13:34.056 nvme0n4: ios=739/1024, merge=0/0, ticks=741/351, in_queue=1092, util=96.25% 00:13:34.056 08:49:23 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t randwrite -r 1 -v 00:13:34.056 [global] 00:13:34.056 thread=1 00:13:34.056 invalidate=1 00:13:34.056 rw=randwrite 00:13:34.056 time_based=1 00:13:34.056 runtime=1 00:13:34.056 ioengine=libaio 00:13:34.056 direct=1 00:13:34.056 bs=4096 00:13:34.056 iodepth=1 00:13:34.056 norandommap=0 00:13:34.056 numjobs=1 00:13:34.056 00:13:34.056 verify_dump=1 00:13:34.056 verify_backlog=512 00:13:34.056 verify_state_save=0 00:13:34.056 do_verify=1 00:13:34.056 verify=crc32c-intel 00:13:34.056 [job0] 00:13:34.056 filename=/dev/nvme0n1 00:13:34.056 [job1] 00:13:34.056 filename=/dev/nvme0n2 00:13:34.056 [job2] 00:13:34.056 filename=/dev/nvme0n3 00:13:34.056 [job3] 00:13:34.056 filename=/dev/nvme0n4 00:13:34.056 Could not set queue depth (nvme0n1) 00:13:34.056 Could not set queue depth (nvme0n2) 00:13:34.056 Could not set queue depth (nvme0n3) 00:13:34.056 Could not set queue depth (nvme0n4) 00:13:34.315 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:34.315 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:34.315 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:34.315 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:34.315 fio-3.35 00:13:34.315 Starting 4 threads 00:13:35.695 00:13:35.695 job0: (groupid=0, jobs=1): err= 0: pid=1754488: Wed Nov 6 08:49:25 2024 00:13:35.695 read: IOPS=382, BW=1531KiB/s (1568kB/s)(1592KiB/1040msec) 00:13:35.695 slat (nsec): min=7110, max=47331, avg=23643.76, stdev=8714.52 00:13:35.695 clat (usec): min=350, max=44953, avg=1812.06, stdev=6741.28 00:13:35.695 lat (usec): min=377, max=44980, avg=1835.70, stdev=6741.47 00:13:35.695 clat percentiles (usec): 00:13:35.695 | 1.00th=[ 396], 5.00th=[ 449], 10.00th=[ 469], 20.00th=[ 510], 00:13:35.695 | 30.00th=[ 537], 40.00th=[ 553], 50.00th=[ 570], 60.00th=[ 586], 00:13:35.695 | 70.00th=[ 750], 80.00th=[ 1004], 90.00th=[ 1057], 95.00th=[ 1156], 00:13:35.695 | 99.00th=[41681], 99.50th=[42206], 99.90th=[44827], 99.95th=[44827], 00:13:35.695 | 99.99th=[44827] 00:13:35.695 write: IOPS=492, BW=1969KiB/s (2016kB/s)(2048KiB/1040msec); 0 zone resets 00:13:35.695 slat (nsec): min=6103, max=62966, avg=28949.05, stdev=10877.67 00:13:35.695 clat (usec): min=155, max=984, avg=555.75, stdev=153.87 00:13:35.695 lat (usec): min=188, max=1003, avg=584.70, stdev=155.96 00:13:35.695 clat percentiles (usec): 00:13:35.695 | 1.00th=[ 225], 5.00th=[ 281], 10.00th=[ 347], 20.00th=[ 424], 00:13:35.695 | 30.00th=[ 474], 40.00th=[ 529], 50.00th=[ 562], 60.00th=[ 611], 00:13:35.695 | 70.00th=[ 644], 80.00th=[ 685], 90.00th=[ 742], 95.00th=[ 799], 00:13:35.695 | 99.00th=[ 922], 99.50th=[ 963], 99.90th=[ 988], 99.95th=[ 988], 00:13:35.695 | 99.99th=[ 988] 00:13:35.695 bw ( KiB/s): min= 4096, max= 4096, per=44.28%, avg=4096.00, stdev= 0.00, samples=1 00:13:35.695 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:35.695 lat (usec) : 250=1.65%, 500=26.15%, 750=53.85%, 1000=9.23% 00:13:35.695 lat (msec) : 2=7.91%, 50=1.21% 00:13:35.695 cpu : usr=1.35%, sys=2.98%, ctx=914, majf=0, minf=1 00:13:35.695 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:35.695 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:35.695 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:35.695 issued rwts: total=398,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:35.695 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:35.695 job1: (groupid=0, jobs=1): err= 0: pid=1754489: Wed Nov 6 08:49:25 2024 00:13:35.695 read: IOPS=202, BW=809KiB/s (829kB/s)(844KiB/1043msec) 00:13:35.695 slat (nsec): min=7920, max=48248, avg=26582.36, stdev=4039.90 00:13:35.695 clat (usec): min=705, max=42981, avg=3355.61, stdev=9517.63 00:13:35.695 lat (usec): min=731, max=43007, avg=3382.19, stdev=9517.40 00:13:35.695 clat percentiles (usec): 00:13:35.695 | 1.00th=[ 758], 5.00th=[ 848], 10.00th=[ 889], 20.00th=[ 930], 00:13:35.695 | 30.00th=[ 971], 40.00th=[ 1004], 50.00th=[ 1029], 60.00th=[ 1057], 00:13:35.695 | 70.00th=[ 1074], 80.00th=[ 1106], 90.00th=[ 1221], 95.00th=[41681], 00:13:35.695 | 99.00th=[42206], 99.50th=[42730], 99.90th=[42730], 99.95th=[42730], 00:13:35.695 | 99.99th=[42730] 00:13:35.695 write: IOPS=490, BW=1964KiB/s (2011kB/s)(2048KiB/1043msec); 0 zone resets 00:13:35.695 slat (nsec): min=5319, max=52628, avg=28753.69, stdev=9403.24 00:13:35.695 clat (usec): min=193, max=1010, avg=603.72, stdev=131.35 00:13:35.695 lat (usec): min=204, max=1043, avg=632.47, stdev=134.35 00:13:35.695 clat percentiles (usec): 00:13:35.695 | 1.00th=[ 302], 5.00th=[ 383], 10.00th=[ 437], 20.00th=[ 494], 00:13:35.695 | 30.00th=[ 537], 40.00th=[ 570], 50.00th=[ 611], 60.00th=[ 644], 00:13:35.695 | 70.00th=[ 676], 80.00th=[ 709], 90.00th=[ 758], 95.00th=[ 816], 00:13:35.695 | 99.00th=[ 930], 99.50th=[ 955], 99.90th=[ 1012], 99.95th=[ 1012], 00:13:35.695 | 99.99th=[ 1012] 00:13:35.695 bw ( KiB/s): min= 4096, max= 4096, per=44.28%, avg=4096.00, stdev= 0.00, samples=1 00:13:35.695 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:35.695 lat (usec) : 250=0.28%, 500=14.94%, 750=47.58%, 1000=18.81% 00:13:35.695 lat (msec) : 2=16.74%, 50=1.66% 00:13:35.695 cpu : usr=1.82%, sys=2.21%, ctx=723, majf=0, minf=2 00:13:35.695 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:35.695 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:35.695 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:35.695 issued rwts: total=211,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:35.695 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:35.695 job2: (groupid=0, jobs=1): err= 0: pid=1754490: Wed Nov 6 08:49:25 2024 00:13:35.695 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:13:35.695 slat (nsec): min=7112, max=62245, avg=22984.86, stdev=8835.28 00:13:35.695 clat (usec): min=149, max=42992, avg=1121.98, stdev=4533.68 00:13:35.695 lat (usec): min=157, max=43019, avg=1144.96, stdev=4534.00 00:13:35.695 clat percentiles (usec): 00:13:35.695 | 1.00th=[ 351], 5.00th=[ 420], 10.00th=[ 453], 20.00th=[ 498], 00:13:35.695 | 30.00th=[ 523], 40.00th=[ 537], 50.00th=[ 553], 60.00th=[ 578], 00:13:35.695 | 70.00th=[ 619], 80.00th=[ 676], 90.00th=[ 963], 95.00th=[ 1004], 00:13:35.695 | 99.00th=[41681], 99.50th=[42206], 99.90th=[43254], 99.95th=[43254], 00:13:35.695 | 99.99th=[43254] 00:13:35.695 write: IOPS=875, BW=3500KiB/s (3585kB/s)(3504KiB/1001msec); 0 zone resets 00:13:35.695 slat (nsec): min=9808, max=67413, avg=28028.48, stdev=11215.34 00:13:35.695 clat (usec): min=101, max=820, avg=429.30, stdev=173.91 00:13:35.695 lat (usec): min=112, max=854, avg=457.33, stdev=176.79 00:13:35.695 clat percentiles (usec): 00:13:35.695 | 1.00th=[ 116], 5.00th=[ 147], 10.00th=[ 237], 20.00th=[ 273], 00:13:35.695 | 30.00th=[ 302], 40.00th=[ 347], 50.00th=[ 400], 60.00th=[ 465], 00:13:35.695 | 70.00th=[ 570], 80.00th=[ 611], 90.00th=[ 676], 95.00th=[ 709], 00:13:35.695 | 99.00th=[ 766], 99.50th=[ 783], 99.90th=[ 824], 99.95th=[ 824], 00:13:35.695 | 99.99th=[ 824] 00:13:35.695 bw ( KiB/s): min= 4104, max= 4104, per=44.37%, avg=4104.00, stdev= 0.00, samples=1 00:13:35.696 iops : min= 1026, max= 1026, avg=1026.00, stdev= 0.00, samples=1 00:13:35.696 lat (usec) : 250=7.85%, 500=39.77%, 750=45.97%, 1000=4.32% 00:13:35.696 lat (msec) : 2=1.59%, 20=0.07%, 50=0.43% 00:13:35.696 cpu : usr=2.20%, sys=3.40%, ctx=1389, majf=0, minf=1 00:13:35.696 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:35.696 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:35.696 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:35.696 issued rwts: total=512,876,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:35.696 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:35.696 job3: (groupid=0, jobs=1): err= 0: pid=1754491: Wed Nov 6 08:49:25 2024 00:13:35.696 read: IOPS=18, BW=74.4KiB/s (76.2kB/s)(76.0KiB/1021msec) 00:13:35.696 slat (nsec): min=26784, max=27624, avg=27160.42, stdev=256.22 00:13:35.696 clat (usec): min=40882, max=42995, avg=41818.65, stdev=664.14 00:13:35.696 lat (usec): min=40909, max=43022, avg=41845.81, stdev=664.10 00:13:35.696 clat percentiles (usec): 00:13:35.696 | 1.00th=[40633], 5.00th=[40633], 10.00th=[41157], 20.00th=[41157], 00:13:35.696 | 30.00th=[41157], 40.00th=[41681], 50.00th=[42206], 60.00th=[42206], 00:13:35.696 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42730], 95.00th=[43254], 00:13:35.696 | 99.00th=[43254], 99.50th=[43254], 99.90th=[43254], 99.95th=[43254], 00:13:35.696 | 99.99th=[43254] 00:13:35.696 write: IOPS=501, BW=2006KiB/s (2054kB/s)(2048KiB/1021msec); 0 zone resets 00:13:35.696 slat (nsec): min=6374, max=47676, avg=14467.06, stdev=11436.38 00:13:35.696 clat (usec): min=208, max=815, avg=414.39, stdev=112.82 00:13:35.696 lat (usec): min=222, max=823, avg=428.86, stdev=115.56 00:13:35.696 clat percentiles (usec): 00:13:35.696 | 1.00th=[ 221], 5.00th=[ 258], 10.00th=[ 273], 20.00th=[ 302], 00:13:35.696 | 30.00th=[ 351], 40.00th=[ 375], 50.00th=[ 408], 60.00th=[ 433], 00:13:35.696 | 70.00th=[ 474], 80.00th=[ 510], 90.00th=[ 570], 95.00th=[ 619], 00:13:35.696 | 99.00th=[ 693], 99.50th=[ 734], 99.90th=[ 816], 99.95th=[ 816], 00:13:35.696 | 99.99th=[ 816] 00:13:35.696 bw ( KiB/s): min= 4096, max= 4096, per=44.28%, avg=4096.00, stdev= 0.00, samples=1 00:13:35.696 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:13:35.696 lat (usec) : 250=2.82%, 500=71.75%, 750=21.47%, 1000=0.38% 00:13:35.696 lat (msec) : 50=3.58% 00:13:35.696 cpu : usr=0.59%, sys=0.59%, ctx=533, majf=0, minf=1 00:13:35.696 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:35.696 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:35.696 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:35.696 issued rwts: total=19,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:35.696 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:35.696 00:13:35.696 Run status group 0 (all jobs): 00:13:35.696 READ: bw=4372KiB/s (4477kB/s), 74.4KiB/s-2046KiB/s (76.2kB/s-2095kB/s), io=4560KiB (4669kB), run=1001-1043msec 00:13:35.696 WRITE: bw=9250KiB/s (9472kB/s), 1964KiB/s-3500KiB/s (2011kB/s-3585kB/s), io=9648KiB (9880kB), run=1001-1043msec 00:13:35.696 00:13:35.696 Disk stats (read/write): 00:13:35.696 nvme0n1: ios=444/512, merge=0/0, ticks=606/246, in_queue=852, util=84.47% 00:13:35.696 nvme0n2: ios=256/512, merge=0/0, ticks=545/248, in_queue=793, util=89.30% 00:13:35.696 nvme0n3: ios=540/621, merge=0/0, ticks=1453/205, in_queue=1658, util=92.50% 00:13:35.696 nvme0n4: ios=37/512, merge=0/0, ticks=1458/203, in_queue=1661, util=93.91% 00:13:35.696 08:49:25 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t write -r 1 -v 00:13:35.696 [global] 00:13:35.696 thread=1 00:13:35.696 invalidate=1 00:13:35.696 rw=write 00:13:35.696 time_based=1 00:13:35.696 runtime=1 00:13:35.696 ioengine=libaio 00:13:35.696 direct=1 00:13:35.696 bs=4096 00:13:35.696 iodepth=128 00:13:35.696 norandommap=0 00:13:35.696 numjobs=1 00:13:35.696 00:13:35.696 verify_dump=1 00:13:35.696 verify_backlog=512 00:13:35.696 verify_state_save=0 00:13:35.696 do_verify=1 00:13:35.696 verify=crc32c-intel 00:13:35.696 [job0] 00:13:35.696 filename=/dev/nvme0n1 00:13:35.696 [job1] 00:13:35.696 filename=/dev/nvme0n2 00:13:35.696 [job2] 00:13:35.696 filename=/dev/nvme0n3 00:13:35.696 [job3] 00:13:35.696 filename=/dev/nvme0n4 00:13:35.696 Could not set queue depth (nvme0n1) 00:13:35.696 Could not set queue depth (nvme0n2) 00:13:35.696 Could not set queue depth (nvme0n3) 00:13:35.696 Could not set queue depth (nvme0n4) 00:13:35.955 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:35.955 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:35.955 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:35.955 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:35.955 fio-3.35 00:13:35.955 Starting 4 threads 00:13:37.337 00:13:37.337 job0: (groupid=0, jobs=1): err= 0: pid=1755013: Wed Nov 6 08:49:27 2024 00:13:37.337 read: IOPS=4838, BW=18.9MiB/s (19.8MB/s)(19.0MiB/1004msec) 00:13:37.337 slat (nsec): min=933, max=46215k, avg=112749.03, stdev=993701.70 00:13:37.337 clat (usec): min=1026, max=67415, avg=14319.64, stdev=10509.71 00:13:37.337 lat (usec): min=3359, max=67421, avg=14432.38, stdev=10559.15 00:13:37.337 clat percentiles (usec): 00:13:37.337 | 1.00th=[ 5342], 5.00th=[ 7570], 10.00th=[ 8455], 20.00th=[ 9241], 00:13:37.337 | 30.00th=[ 9765], 40.00th=[10290], 50.00th=[10683], 60.00th=[11207], 00:13:37.337 | 70.00th=[13042], 80.00th=[17957], 90.00th=[22414], 95.00th=[40109], 00:13:37.337 | 99.00th=[65799], 99.50th=[67634], 99.90th=[67634], 99.95th=[67634], 00:13:37.337 | 99.99th=[67634] 00:13:37.337 write: IOPS=5099, BW=19.9MiB/s (20.9MB/s)(20.0MiB/1004msec); 0 zone resets 00:13:37.337 slat (nsec): min=1596, max=9211.6k, avg=79503.69, stdev=489518.33 00:13:37.337 clat (usec): min=1484, max=25920, avg=11146.05, stdev=3614.56 00:13:37.337 lat (usec): min=1493, max=25930, avg=11225.55, stdev=3628.70 00:13:37.337 clat percentiles (usec): 00:13:37.337 | 1.00th=[ 3949], 5.00th=[ 6783], 10.00th=[ 7898], 20.00th=[ 9372], 00:13:37.337 | 30.00th=[ 9765], 40.00th=[10028], 50.00th=[10290], 60.00th=[10683], 00:13:37.337 | 70.00th=[11207], 80.00th=[11731], 90.00th=[16712], 95.00th=[19006], 00:13:37.337 | 99.00th=[24249], 99.50th=[25822], 99.90th=[25822], 99.95th=[25822], 00:13:37.337 | 99.99th=[25822] 00:13:37.337 bw ( KiB/s): min=16384, max=24576, per=23.56%, avg=20480.00, stdev=5792.62, samples=2 00:13:37.337 iops : min= 4096, max= 6144, avg=5120.00, stdev=1448.15, samples=2 00:13:37.337 lat (msec) : 2=0.10%, 4=1.07%, 10=34.26%, 20=54.36%, 50=8.94% 00:13:37.337 lat (msec) : 100=1.27% 00:13:37.337 cpu : usr=2.89%, sys=5.58%, ctx=448, majf=0, minf=2 00:13:37.337 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.4% 00:13:37.337 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:37.337 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:37.337 issued rwts: total=4858,5120,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:37.337 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:37.337 job1: (groupid=0, jobs=1): err= 0: pid=1755017: Wed Nov 6 08:49:27 2024 00:13:37.337 read: IOPS=6629, BW=25.9MiB/s (27.2MB/s)(26.0MiB/1004msec) 00:13:37.337 slat (nsec): min=912, max=45717k, avg=71393.36, stdev=723537.52 00:13:37.337 clat (usec): min=3420, max=62186, avg=9594.53, stdev=7393.11 00:13:37.337 lat (usec): min=3426, max=62192, avg=9665.93, stdev=7434.86 00:13:37.337 clat percentiles (usec): 00:13:37.337 | 1.00th=[ 4080], 5.00th=[ 5145], 10.00th=[ 5800], 20.00th=[ 6456], 00:13:37.337 | 30.00th=[ 6915], 40.00th=[ 7373], 50.00th=[ 7767], 60.00th=[ 8160], 00:13:37.338 | 70.00th=[ 8717], 80.00th=[10290], 90.00th=[14222], 95.00th=[18220], 00:13:37.338 | 99.00th=[54264], 99.50th=[56886], 99.90th=[62129], 99.95th=[62129], 00:13:37.338 | 99.99th=[62129] 00:13:37.338 write: IOPS=6867, BW=26.8MiB/s (28.1MB/s)(26.9MiB/1004msec); 0 zone resets 00:13:37.338 slat (nsec): min=1616, max=14588k, avg=67171.12, stdev=488691.40 00:13:37.338 clat (usec): min=1258, max=40939, avg=9191.81, stdev=5253.76 00:13:37.338 lat (usec): min=1269, max=40947, avg=9258.99, stdev=5291.42 00:13:37.338 clat percentiles (usec): 00:13:37.338 | 1.00th=[ 3195], 5.00th=[ 4228], 10.00th=[ 4883], 20.00th=[ 6063], 00:13:37.338 | 30.00th=[ 6652], 40.00th=[ 7439], 50.00th=[ 7963], 60.00th=[ 8586], 00:13:37.338 | 70.00th=[ 9241], 80.00th=[10159], 90.00th=[15795], 95.00th=[20317], 00:13:37.338 | 99.00th=[27919], 99.50th=[36439], 99.90th=[41157], 99.95th=[41157], 00:13:37.338 | 99.99th=[41157] 00:13:37.338 bw ( KiB/s): min=25480, max=28664, per=31.15%, avg=27072.00, stdev=2251.43, samples=2 00:13:37.338 iops : min= 6370, max= 7166, avg=6768.00, stdev=562.86, samples=2 00:13:37.338 lat (msec) : 2=0.12%, 4=2.44%, 10=75.44%, 20=17.45%, 50=3.85% 00:13:37.338 lat (msec) : 100=0.71% 00:13:37.338 cpu : usr=5.08%, sys=7.28%, ctx=475, majf=0, minf=1 00:13:37.338 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.5% 00:13:37.338 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:37.338 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:37.338 issued rwts: total=6656,6895,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:37.338 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:37.338 job2: (groupid=0, jobs=1): err= 0: pid=1755021: Wed Nov 6 08:49:27 2024 00:13:37.338 read: IOPS=4594, BW=17.9MiB/s (18.8MB/s)(18.0MiB/1003msec) 00:13:37.338 slat (nsec): min=951, max=44330k, avg=108081.93, stdev=893442.57 00:13:37.338 clat (usec): min=2151, max=53893, avg=13805.06, stdev=6953.47 00:13:37.338 lat (usec): min=2156, max=53895, avg=13913.15, stdev=7007.99 00:13:37.338 clat percentiles (usec): 00:13:37.338 | 1.00th=[ 2999], 5.00th=[ 6718], 10.00th=[ 8979], 20.00th=[ 9896], 00:13:37.338 | 30.00th=[10290], 40.00th=[11207], 50.00th=[11994], 60.00th=[13042], 00:13:37.338 | 70.00th=[15533], 80.00th=[16188], 90.00th=[20841], 95.00th=[26608], 00:13:37.338 | 99.00th=[46400], 99.50th=[52691], 99.90th=[53740], 99.95th=[53740], 00:13:37.338 | 99.99th=[53740] 00:13:37.338 write: IOPS=5043, BW=19.7MiB/s (20.7MB/s)(19.8MiB/1003msec); 0 zone resets 00:13:37.338 slat (nsec): min=1646, max=16023k, avg=94554.05, stdev=580015.69 00:13:37.338 clat (usec): min=478, max=53887, avg=12489.75, stdev=7506.92 00:13:37.338 lat (usec): min=4394, max=53888, avg=12584.31, stdev=7526.28 00:13:37.338 clat percentiles (usec): 00:13:37.338 | 1.00th=[ 4883], 5.00th=[ 6652], 10.00th=[ 7504], 20.00th=[ 8717], 00:13:37.338 | 30.00th=[ 9503], 40.00th=[10028], 50.00th=[10945], 60.00th=[11338], 00:13:37.338 | 70.00th=[12649], 80.00th=[14091], 90.00th=[16450], 95.00th=[21103], 00:13:37.338 | 99.00th=[52691], 99.50th=[53740], 99.90th=[53740], 99.95th=[53740], 00:13:37.338 | 99.99th=[53740] 00:13:37.338 bw ( KiB/s): min=19136, max=20312, per=22.69%, avg=19724.00, stdev=831.56, samples=2 00:13:37.338 iops : min= 4784, max= 5078, avg=4931.00, stdev=207.89, samples=2 00:13:37.338 lat (usec) : 500=0.01% 00:13:37.338 lat (msec) : 4=1.09%, 10=29.77%, 20=61.15%, 50=6.60%, 100=1.39% 00:13:37.338 cpu : usr=2.30%, sys=4.49%, ctx=471, majf=0, minf=1 00:13:37.338 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.3% 00:13:37.338 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:37.338 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:37.338 issued rwts: total=4608,5059,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:37.338 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:37.338 job3: (groupid=0, jobs=1): err= 0: pid=1755022: Wed Nov 6 08:49:27 2024 00:13:37.338 read: IOPS=4960, BW=19.4MiB/s (20.3MB/s)(20.2MiB/1045msec) 00:13:37.338 slat (nsec): min=959, max=10295k, avg=97487.37, stdev=648575.11 00:13:37.338 clat (usec): min=3787, max=56751, avg=12156.17, stdev=6190.86 00:13:37.338 lat (usec): min=3796, max=56765, avg=12253.65, stdev=6234.46 00:13:37.338 clat percentiles (usec): 00:13:37.338 | 1.00th=[ 6259], 5.00th=[ 7635], 10.00th=[ 7832], 20.00th=[ 8586], 00:13:37.338 | 30.00th=[ 8979], 40.00th=[ 9241], 50.00th=[10028], 60.00th=[11600], 00:13:37.338 | 70.00th=[12911], 80.00th=[15533], 90.00th=[17433], 95.00th=[21103], 00:13:37.338 | 99.00th=[49546], 99.50th=[56886], 99.90th=[56886], 99.95th=[56886], 00:13:37.338 | 99.99th=[56886] 00:13:37.338 write: IOPS=5389, BW=21.1MiB/s (22.1MB/s)(22.0MiB/1045msec); 0 zone resets 00:13:37.338 slat (nsec): min=1621, max=11756k, avg=82860.68, stdev=435862.65 00:13:37.338 clat (usec): min=688, max=65756, avg=12332.99, stdev=7909.12 00:13:37.338 lat (usec): min=696, max=65761, avg=12415.85, stdev=7931.53 00:13:37.338 clat percentiles (usec): 00:13:37.338 | 1.00th=[ 1516], 5.00th=[ 4359], 10.00th=[ 6063], 20.00th=[ 7767], 00:13:37.338 | 30.00th=[ 8455], 40.00th=[ 8979], 50.00th=[ 9503], 60.00th=[12387], 00:13:37.338 | 70.00th=[14877], 80.00th=[17171], 90.00th=[19530], 95.00th=[21890], 00:13:37.338 | 99.00th=[61080], 99.50th=[65799], 99.90th=[65799], 99.95th=[65799], 00:13:37.338 | 99.99th=[65799] 00:13:37.338 bw ( KiB/s): min=19920, max=24624, per=25.63%, avg=22272.00, stdev=3326.23, samples=2 00:13:37.338 iops : min= 4980, max= 6156, avg=5568.00, stdev=831.56, samples=2 00:13:37.338 lat (usec) : 750=0.03% 00:13:37.338 lat (msec) : 2=0.66%, 4=1.49%, 10=49.33%, 20=41.14%, 50=6.48% 00:13:37.338 lat (msec) : 100=0.87% 00:13:37.338 cpu : usr=2.30%, sys=5.56%, ctx=625, majf=0, minf=1 00:13:37.338 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:13:37.338 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:37.338 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:37.338 issued rwts: total=5184,5632,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:37.338 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:37.338 00:13:37.338 Run status group 0 (all jobs): 00:13:37.338 READ: bw=79.6MiB/s (83.5MB/s), 17.9MiB/s-25.9MiB/s (18.8MB/s-27.2MB/s), io=83.2MiB (87.3MB), run=1003-1045msec 00:13:37.338 WRITE: bw=84.9MiB/s (89.0MB/s), 19.7MiB/s-26.8MiB/s (20.7MB/s-28.1MB/s), io=88.7MiB (93.0MB), run=1003-1045msec 00:13:37.338 00:13:37.338 Disk stats (read/write): 00:13:37.338 nvme0n1: ios=4142/4240, merge=0/0, ticks=21339/19465, in_queue=40804, util=86.07% 00:13:37.338 nvme0n2: ios=5425/5632, merge=0/0, ticks=27391/26476, in_queue=53867, util=86.95% 00:13:37.338 nvme0n3: ios=4100/4096, merge=0/0, ticks=21421/19206, in_queue=40627, util=92.30% 00:13:37.338 nvme0n4: ios=4665/4959, merge=0/0, ticks=43853/46296, in_queue=90149, util=96.91% 00:13:37.338 08:49:27 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t randwrite -r 1 -v 00:13:37.338 [global] 00:13:37.338 thread=1 00:13:37.338 invalidate=1 00:13:37.338 rw=randwrite 00:13:37.338 time_based=1 00:13:37.338 runtime=1 00:13:37.338 ioengine=libaio 00:13:37.338 direct=1 00:13:37.338 bs=4096 00:13:37.338 iodepth=128 00:13:37.338 norandommap=0 00:13:37.338 numjobs=1 00:13:37.338 00:13:37.338 verify_dump=1 00:13:37.338 verify_backlog=512 00:13:37.338 verify_state_save=0 00:13:37.338 do_verify=1 00:13:37.338 verify=crc32c-intel 00:13:37.338 [job0] 00:13:37.338 filename=/dev/nvme0n1 00:13:37.338 [job1] 00:13:37.338 filename=/dev/nvme0n2 00:13:37.338 [job2] 00:13:37.338 filename=/dev/nvme0n3 00:13:37.338 [job3] 00:13:37.338 filename=/dev/nvme0n4 00:13:37.338 Could not set queue depth (nvme0n1) 00:13:37.338 Could not set queue depth (nvme0n2) 00:13:37.338 Could not set queue depth (nvme0n3) 00:13:37.338 Could not set queue depth (nvme0n4) 00:13:37.906 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:37.906 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:37.906 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:37.906 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:13:37.906 fio-3.35 00:13:37.906 Starting 4 threads 00:13:39.301 00:13:39.301 job0: (groupid=0, jobs=1): err= 0: pid=1755542: Wed Nov 6 08:49:28 2024 00:13:39.301 read: IOPS=5615, BW=21.9MiB/s (23.0MB/s)(22.0MiB/1003msec) 00:13:39.301 slat (nsec): min=888, max=13397k, avg=84474.97, stdev=503069.90 00:13:39.301 clat (usec): min=5220, max=47812, avg=10837.15, stdev=4852.71 00:13:39.301 lat (usec): min=5227, max=47837, avg=10921.62, stdev=4890.19 00:13:39.301 clat percentiles (usec): 00:13:39.301 | 1.00th=[ 7242], 5.00th=[ 7832], 10.00th=[ 8291], 20.00th=[ 8848], 00:13:39.301 | 30.00th=[ 9110], 40.00th=[ 9241], 50.00th=[ 9503], 60.00th=[ 9765], 00:13:39.301 | 70.00th=[10028], 80.00th=[10421], 90.00th=[15270], 95.00th=[20055], 00:13:39.301 | 99.00th=[36439], 99.50th=[40109], 99.90th=[40109], 99.95th=[40109], 00:13:39.301 | 99.99th=[47973] 00:13:39.301 write: IOPS=5710, BW=22.3MiB/s (23.4MB/s)(22.4MiB/1003msec); 0 zone resets 00:13:39.301 slat (nsec): min=1477, max=23486k, avg=87953.57, stdev=680966.06 00:13:39.301 clat (usec): min=588, max=57291, avg=11520.94, stdev=8100.30 00:13:39.301 lat (usec): min=2421, max=57321, avg=11608.89, stdev=8163.39 00:13:39.301 clat percentiles (usec): 00:13:39.301 | 1.00th=[ 5145], 5.00th=[ 6915], 10.00th=[ 7177], 20.00th=[ 7439], 00:13:39.301 | 30.00th=[ 7898], 40.00th=[ 8586], 50.00th=[ 8848], 60.00th=[ 9241], 00:13:39.301 | 70.00th=[ 9634], 80.00th=[10421], 90.00th=[20055], 95.00th=[32637], 00:13:39.301 | 99.00th=[44827], 99.50th=[54264], 99.90th=[54264], 99.95th=[54264], 00:13:39.301 | 99.99th=[57410] 00:13:39.301 bw ( KiB/s): min=19928, max=25088, per=27.83%, avg=22508.00, stdev=3648.67, samples=2 00:13:39.301 iops : min= 4982, max= 6272, avg=5627.00, stdev=912.17, samples=2 00:13:39.301 lat (usec) : 750=0.01% 00:13:39.301 lat (msec) : 4=0.28%, 10=71.73%, 20=20.33%, 50=7.36%, 100=0.30% 00:13:39.301 cpu : usr=2.59%, sys=3.79%, ctx=620, majf=0, minf=2 00:13:39.301 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:13:39.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:39.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:39.301 issued rwts: total=5632,5728,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:39.301 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:39.301 job1: (groupid=0, jobs=1): err= 0: pid=1755543: Wed Nov 6 08:49:28 2024 00:13:39.301 read: IOPS=2856, BW=11.2MiB/s (11.7MB/s)(11.6MiB/1043msec) 00:13:39.301 slat (nsec): min=1381, max=19180k, avg=161367.30, stdev=1032471.15 00:13:39.301 clat (usec): min=6387, max=73160, avg=24802.56, stdev=15326.66 00:13:39.301 lat (usec): min=6392, max=73186, avg=24963.92, stdev=15336.83 00:13:39.301 clat percentiles (usec): 00:13:39.301 | 1.00th=[ 6456], 5.00th=[ 7635], 10.00th=[ 7963], 20.00th=[14484], 00:13:39.301 | 30.00th=[18220], 40.00th=[21103], 50.00th=[22676], 60.00th=[23462], 00:13:39.301 | 70.00th=[24773], 80.00th=[28181], 90.00th=[50594], 95.00th=[63177], 00:13:39.301 | 99.00th=[72877], 99.50th=[72877], 99.90th=[72877], 99.95th=[72877], 00:13:39.301 | 99.99th=[72877] 00:13:39.301 write: IOPS=2945, BW=11.5MiB/s (12.1MB/s)(12.0MiB/1043msec); 0 zone resets 00:13:39.301 slat (nsec): min=1595, max=37077k, avg=162639.33, stdev=1281453.79 00:13:39.301 clat (usec): min=4733, max=46831, avg=18423.99, stdev=8692.98 00:13:39.301 lat (usec): min=4740, max=63401, avg=18586.63, stdev=8782.06 00:13:39.301 clat percentiles (usec): 00:13:39.301 | 1.00th=[ 7046], 5.00th=[ 7439], 10.00th=[ 7570], 20.00th=[ 7898], 00:13:39.301 | 30.00th=[13173], 40.00th=[16909], 50.00th=[18744], 60.00th=[20055], 00:13:39.301 | 70.00th=[22676], 80.00th=[24773], 90.00th=[28705], 95.00th=[31327], 00:13:39.301 | 99.00th=[45876], 99.50th=[46924], 99.90th=[46924], 99.95th=[46924], 00:13:39.301 | 99.99th=[46924] 00:13:39.301 bw ( KiB/s): min=10488, max=14059, per=15.18%, avg=12273.50, stdev=2525.08, samples=2 00:13:39.301 iops : min= 2622, max= 3514, avg=3068.00, stdev=630.74, samples=2 00:13:39.301 lat (msec) : 10=22.01%, 20=25.80%, 50=47.20%, 100=4.99% 00:13:39.301 cpu : usr=2.69%, sys=3.07%, ctx=203, majf=0, minf=2 00:13:39.301 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.3%, 32=0.5%, >=64=99.0% 00:13:39.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:39.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:39.301 issued rwts: total=2979,3072,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:39.301 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:39.301 job2: (groupid=0, jobs=1): err= 0: pid=1755544: Wed Nov 6 08:49:28 2024 00:13:39.301 read: IOPS=4442, BW=17.4MiB/s (18.2MB/s)(17.4MiB/1003msec) 00:13:39.301 slat (nsec): min=1081, max=12015k, avg=109485.82, stdev=789579.95 00:13:39.301 clat (usec): min=1941, max=28991, avg=14470.66, stdev=3990.13 00:13:39.301 lat (usec): min=1949, max=29016, avg=14580.15, stdev=4064.01 00:13:39.301 clat percentiles (usec): 00:13:39.301 | 1.00th=[ 5932], 5.00th=[ 7898], 10.00th=[ 8979], 20.00th=[10552], 00:13:39.301 | 30.00th=[11469], 40.00th=[13960], 50.00th=[15401], 60.00th=[16057], 00:13:39.301 | 70.00th=[16712], 80.00th=[17957], 90.00th=[19530], 95.00th=[20055], 00:13:39.301 | 99.00th=[22938], 99.50th=[23200], 99.90th=[26084], 99.95th=[28967], 00:13:39.301 | 99.99th=[28967] 00:13:39.301 write: IOPS=4594, BW=17.9MiB/s (18.8MB/s)(18.0MiB/1003msec); 0 zone resets 00:13:39.301 slat (nsec): min=1634, max=9271.1k, avg=75538.56, stdev=504558.05 00:13:39.301 clat (usec): min=1149, max=58417, avg=13608.52, stdev=8546.47 00:13:39.301 lat (usec): min=1158, max=58420, avg=13684.06, stdev=8592.73 00:13:39.301 clat percentiles (usec): 00:13:39.301 | 1.00th=[ 2802], 5.00th=[ 5407], 10.00th=[ 6456], 20.00th=[ 8160], 00:13:39.301 | 30.00th=[ 8717], 40.00th=[10290], 50.00th=[11076], 60.00th=[12387], 00:13:39.301 | 70.00th=[15533], 80.00th=[17433], 90.00th=[24511], 95.00th=[28443], 00:13:39.301 | 99.00th=[50070], 99.50th=[52167], 99.90th=[53740], 99.95th=[53740], 00:13:39.301 | 99.99th=[58459] 00:13:39.301 bw ( KiB/s): min=18123, max=18704, per=22.77%, avg=18413.50, stdev=410.83, samples=2 00:13:39.301 iops : min= 4530, max= 4676, avg=4603.00, stdev=103.24, samples=2 00:13:39.301 lat (msec) : 2=0.26%, 4=1.35%, 10=26.74%, 20=61.61%, 50=9.58% 00:13:39.301 lat (msec) : 100=0.46% 00:13:39.301 cpu : usr=3.89%, sys=5.29%, ctx=344, majf=0, minf=1 00:13:39.301 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:13:39.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:39.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:39.301 issued rwts: total=4456,4608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:39.301 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:39.301 job3: (groupid=0, jobs=1): err= 0: pid=1755545: Wed Nov 6 08:49:28 2024 00:13:39.301 read: IOPS=7633, BW=29.8MiB/s (31.3MB/s)(30.0MiB/1005msec) 00:13:39.301 slat (nsec): min=973, max=8291.1k, avg=65037.85, stdev=492396.84 00:13:39.301 clat (usec): min=907, max=17302, avg=8599.55, stdev=1959.17 00:13:39.301 lat (usec): min=2732, max=17328, avg=8664.58, stdev=1992.66 00:13:39.301 clat percentiles (usec): 00:13:39.301 | 1.00th=[ 4883], 5.00th=[ 6063], 10.00th=[ 6587], 20.00th=[ 7308], 00:13:39.301 | 30.00th=[ 7635], 40.00th=[ 7832], 50.00th=[ 8160], 60.00th=[ 8586], 00:13:39.301 | 70.00th=[ 9110], 80.00th=[ 9765], 90.00th=[11469], 95.00th=[12649], 00:13:39.301 | 99.00th=[15008], 99.50th=[15795], 99.90th=[16450], 99.95th=[16581], 00:13:39.301 | 99.99th=[17433] 00:13:39.301 write: IOPS=7641, BW=29.9MiB/s (31.3MB/s)(30.0MiB/1005msec); 0 zone resets 00:13:39.301 slat (nsec): min=1675, max=7544.7k, avg=60029.24, stdev=408902.45 00:13:39.301 clat (usec): min=1481, max=23679, avg=7987.04, stdev=2900.81 00:13:39.301 lat (usec): min=1489, max=23689, avg=8047.07, stdev=2929.43 00:13:39.301 clat percentiles (usec): 00:13:39.301 | 1.00th=[ 3195], 5.00th=[ 4555], 10.00th=[ 4883], 20.00th=[ 6194], 00:13:39.301 | 30.00th=[ 6980], 40.00th=[ 7308], 50.00th=[ 7635], 60.00th=[ 7832], 00:13:39.301 | 70.00th=[ 8160], 80.00th=[ 9110], 90.00th=[10552], 95.00th=[13960], 00:13:39.301 | 99.00th=[20055], 99.50th=[20579], 99.90th=[23462], 99.95th=[23462], 00:13:39.301 | 99.99th=[23725] 00:13:39.301 bw ( KiB/s): min=29552, max=31888, per=37.98%, avg=30720.00, stdev=1651.80, samples=2 00:13:39.301 iops : min= 7388, max= 7972, avg=7680.00, stdev=412.95, samples=2 00:13:39.301 lat (usec) : 1000=0.01% 00:13:39.301 lat (msec) : 2=0.04%, 4=1.54%, 10=82.12%, 20=15.80%, 50=0.50% 00:13:39.301 cpu : usr=5.88%, sys=8.17%, ctx=573, majf=0, minf=1 00:13:39.301 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.6% 00:13:39.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:39.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:13:39.301 issued rwts: total=7672,7680,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:39.301 latency : target=0, window=0, percentile=100.00%, depth=128 00:13:39.301 00:13:39.301 Run status group 0 (all jobs): 00:13:39.302 READ: bw=77.7MiB/s (81.4MB/s), 11.2MiB/s-29.8MiB/s (11.7MB/s-31.3MB/s), io=81.0MiB (84.9MB), run=1003-1043msec 00:13:39.302 WRITE: bw=79.0MiB/s (82.8MB/s), 11.5MiB/s-29.9MiB/s (12.1MB/s-31.3MB/s), io=82.4MiB (86.4MB), run=1003-1043msec 00:13:39.302 00:13:39.302 Disk stats (read/write): 00:13:39.302 nvme0n1: ios=4594/4608, merge=0/0, ticks=16681/18240, in_queue=34921, util=87.78% 00:13:39.302 nvme0n2: ios=2398/2560, merge=0/0, ticks=16935/13366, in_queue=30301, util=91.44% 00:13:39.302 nvme0n3: ios=3634/3711, merge=0/0, ticks=33384/43650, in_queue=77034, util=95.36% 00:13:39.302 nvme0n4: ios=6194/6639, merge=0/0, ticks=50182/51002, in_queue=101184, util=97.97% 00:13:39.302 08:49:28 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@55 -- # sync 00:13:39.302 08:49:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@59 -- # fio_pid=1755873 00:13:39.302 08:49:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t read -r 10 00:13:39.302 08:49:29 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@61 -- # sleep 3 00:13:39.302 [global] 00:13:39.302 thread=1 00:13:39.302 invalidate=1 00:13:39.302 rw=read 00:13:39.302 time_based=1 00:13:39.302 runtime=10 00:13:39.302 ioengine=libaio 00:13:39.302 direct=1 00:13:39.302 bs=4096 00:13:39.302 iodepth=1 00:13:39.302 norandommap=1 00:13:39.302 numjobs=1 00:13:39.302 00:13:39.302 [job0] 00:13:39.302 filename=/dev/nvme0n1 00:13:39.302 [job1] 00:13:39.302 filename=/dev/nvme0n2 00:13:39.302 [job2] 00:13:39.302 filename=/dev/nvme0n3 00:13:39.302 [job3] 00:13:39.302 filename=/dev/nvme0n4 00:13:39.302 Could not set queue depth (nvme0n1) 00:13:39.302 Could not set queue depth (nvme0n2) 00:13:39.302 Could not set queue depth (nvme0n3) 00:13:39.302 Could not set queue depth (nvme0n4) 00:13:39.562 job0: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:39.562 job1: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:39.562 job2: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:39.562 job3: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:13:39.562 fio-3.35 00:13:39.562 Starting 4 threads 00:13:42.107 08:49:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete concat0 00:13:42.107 08:49:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete raid0 00:13:42.107 fio: io_u error on file /dev/nvme0n4: Operation not supported: read offset=2269184, buflen=4096 00:13:42.107 fio: pid=1756075, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:13:42.367 fio: io_u error on file /dev/nvme0n3: Operation not supported: read offset=8634368, buflen=4096 00:13:42.367 fio: pid=1756074, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:13:42.367 08:49:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:42.367 08:49:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc0 00:13:42.628 08:49:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:42.628 08:49:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc1 00:13:42.628 fio: io_u error on file /dev/nvme0n1: Operation not supported: read offset=2838528, buflen=4096 00:13:42.628 fio: pid=1756071, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:13:42.888 fio: io_u error on file /dev/nvme0n2: Operation not supported: read offset=3846144, buflen=4096 00:13:42.888 fio: pid=1756073, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:13:42.888 08:49:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:42.888 08:49:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc2 00:13:42.888 00:13:42.888 job0: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=1756071: Wed Nov 6 08:49:32 2024 00:13:42.888 read: IOPS=231, BW=925KiB/s (947kB/s)(2772KiB/2996msec) 00:13:42.888 slat (usec): min=2, max=288, avg=23.57, stdev=13.46 00:13:42.888 clat (usec): min=364, max=42215, avg=4261.88, stdev=11350.63 00:13:42.888 lat (usec): min=373, max=42241, avg=4285.06, stdev=11352.43 00:13:42.888 clat percentiles (usec): 00:13:42.889 | 1.00th=[ 529], 5.00th=[ 619], 10.00th=[ 668], 20.00th=[ 709], 00:13:42.889 | 30.00th=[ 742], 40.00th=[ 758], 50.00th=[ 791], 60.00th=[ 807], 00:13:42.889 | 70.00th=[ 832], 80.00th=[ 865], 90.00th=[ 930], 95.00th=[41157], 00:13:42.889 | 99.00th=[41681], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:13:42.889 | 99.99th=[42206] 00:13:42.889 bw ( KiB/s): min= 96, max= 4784, per=20.07%, avg=1089.60, stdev=2066.75, samples=5 00:13:42.889 iops : min= 24, max= 1196, avg=272.40, stdev=516.69, samples=5 00:13:42.889 lat (usec) : 500=0.72%, 750=33.29%, 1000=57.20% 00:13:42.889 lat (msec) : 50=8.65% 00:13:42.889 cpu : usr=0.17%, sys=0.67%, ctx=696, majf=0, minf=2 00:13:42.889 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:42.889 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:42.889 complete : 0=0.1%, 4=99.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:42.889 issued rwts: total=694,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:42.889 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:42.889 job1: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=1756073: Wed Nov 6 08:49:32 2024 00:13:42.889 read: IOPS=296, BW=1187KiB/s (1215kB/s)(3756KiB/3165msec) 00:13:42.889 slat (usec): min=6, max=18939, avg=64.27, stdev=788.36 00:13:42.889 clat (usec): min=225, max=43125, avg=3276.73, stdev=9540.50 00:13:42.889 lat (usec): min=232, max=43152, avg=3338.47, stdev=9564.55 00:13:42.889 clat percentiles (usec): 00:13:42.889 | 1.00th=[ 404], 5.00th=[ 515], 10.00th=[ 619], 20.00th=[ 766], 00:13:42.889 | 30.00th=[ 840], 40.00th=[ 898], 50.00th=[ 938], 60.00th=[ 971], 00:13:42.889 | 70.00th=[ 1012], 80.00th=[ 1057], 90.00th=[ 1123], 95.00th=[41157], 00:13:42.889 | 99.00th=[42206], 99.50th=[42730], 99.90th=[43254], 99.95th=[43254], 00:13:42.889 | 99.99th=[43254] 00:13:42.889 bw ( KiB/s): min= 96, max= 3431, per=21.25%, avg=1153.17, stdev=1617.84, samples=6 00:13:42.889 iops : min= 24, max= 857, avg=288.17, stdev=404.25, samples=6 00:13:42.889 lat (usec) : 250=0.11%, 500=3.62%, 750=13.72%, 1000=49.89% 00:13:42.889 lat (msec) : 2=26.60%, 20=0.11%, 50=5.85% 00:13:42.889 cpu : usr=0.28%, sys=0.92%, ctx=945, majf=0, minf=2 00:13:42.889 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:42.889 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:42.889 complete : 0=0.1%, 4=99.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:42.889 issued rwts: total=940,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:42.889 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:42.889 job2: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=1756074: Wed Nov 6 08:49:32 2024 00:13:42.889 read: IOPS=755, BW=3019KiB/s (3091kB/s)(8432KiB/2793msec) 00:13:42.889 slat (usec): min=7, max=21453, avg=43.61, stdev=568.20 00:13:42.889 clat (usec): min=248, max=42574, avg=1262.29, stdev=2902.01 00:13:42.889 lat (usec): min=274, max=42600, avg=1305.91, stdev=2955.89 00:13:42.889 clat percentiles (usec): 00:13:42.889 | 1.00th=[ 594], 5.00th=[ 791], 10.00th=[ 865], 20.00th=[ 947], 00:13:42.889 | 30.00th=[ 1004], 40.00th=[ 1045], 50.00th=[ 1090], 60.00th=[ 1123], 00:13:42.889 | 70.00th=[ 1139], 80.00th=[ 1172], 90.00th=[ 1205], 95.00th=[ 1237], 00:13:42.889 | 99.00th=[ 1303], 99.50th=[40633], 99.90th=[41157], 99.95th=[41681], 00:13:42.889 | 99.99th=[42730] 00:13:42.889 bw ( KiB/s): min= 2648, max= 3224, per=55.63%, avg=3019.20, stdev=247.91, samples=5 00:13:42.889 iops : min= 662, max= 806, avg=754.80, stdev=61.98, samples=5 00:13:42.889 lat (usec) : 250=0.05%, 500=0.28%, 750=3.03%, 1000=26.70% 00:13:42.889 lat (msec) : 2=69.37%, 50=0.52% 00:13:42.889 cpu : usr=0.93%, sys=2.22%, ctx=2111, majf=0, minf=2 00:13:42.889 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:42.889 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:42.889 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:42.889 issued rwts: total=2109,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:42.889 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:42.889 job3: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=1756075: Wed Nov 6 08:49:32 2024 00:13:42.889 read: IOPS=212, BW=847KiB/s (867kB/s)(2216KiB/2617msec) 00:13:42.889 slat (nsec): min=6817, max=61524, avg=26891.09, stdev=3812.00 00:13:42.889 clat (usec): min=399, max=43027, avg=4650.48, stdev=11766.86 00:13:42.889 lat (usec): min=426, max=43053, avg=4677.37, stdev=11766.71 00:13:42.889 clat percentiles (usec): 00:13:42.889 | 1.00th=[ 627], 5.00th=[ 742], 10.00th=[ 799], 20.00th=[ 857], 00:13:42.889 | 30.00th=[ 889], 40.00th=[ 914], 50.00th=[ 930], 60.00th=[ 947], 00:13:42.889 | 70.00th=[ 971], 80.00th=[ 996], 90.00th=[ 1123], 95.00th=[41681], 00:13:42.889 | 99.00th=[42206], 99.50th=[42206], 99.90th=[43254], 99.95th=[43254], 00:13:42.889 | 99.99th=[43254] 00:13:42.889 bw ( KiB/s): min= 504, max= 1504, per=16.14%, avg=876.80, stdev=408.06, samples=5 00:13:42.889 iops : min= 126, max= 376, avg=219.20, stdev=102.02, samples=5 00:13:42.889 lat (usec) : 500=0.36%, 750=5.23%, 1000=75.50% 00:13:42.889 lat (msec) : 2=9.55%, 50=9.19% 00:13:42.889 cpu : usr=0.46%, sys=0.76%, ctx=555, majf=0, minf=1 00:13:42.889 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:13:42.889 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:42.889 complete : 0=0.2%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:13:42.889 issued rwts: total=555,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:13:42.889 latency : target=0, window=0, percentile=100.00%, depth=1 00:13:42.889 00:13:42.889 Run status group 0 (all jobs): 00:13:42.889 READ: bw=5427KiB/s (5557kB/s), 847KiB/s-3019KiB/s (867kB/s-3091kB/s), io=16.8MiB (17.6MB), run=2617-3165msec 00:13:42.889 00:13:42.889 Disk stats (read/write): 00:13:42.889 nvme0n1: ios=689/0, merge=0/0, ticks=2771/0, in_queue=2771, util=94.79% 00:13:42.889 nvme0n2: ios=971/0, merge=0/0, ticks=3206/0, in_queue=3206, util=98.51% 00:13:42.889 nvme0n3: ios=1959/0, merge=0/0, ticks=2474/0, in_queue=2474, util=96.03% 00:13:42.889 nvme0n4: ios=553/0, merge=0/0, ticks=2484/0, in_queue=2484, util=96.46% 00:13:42.889 08:49:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:42.889 08:49:32 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc3 00:13:43.150 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:43.150 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc4 00:13:43.410 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:43.410 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc5 00:13:43.410 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:13:43.410 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc6 00:13:43.671 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@69 -- # fio_status=0 00:13:43.671 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@70 -- # wait 1755873 00:13:43.671 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@70 -- # fio_status=4 00:13:43.671 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@72 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:13:43.671 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:13:43.671 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@73 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:13:43.671 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1219 -- # local i=0 00:13:43.671 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:13:43.671 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:43.931 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:13:43.931 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:13:43.931 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1231 -- # return 0 00:13:43.931 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@75 -- # '[' 4 -eq 0 ']' 00:13:43.931 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@80 -- # echo 'nvmf hotplug test: fio failed as expected' 00:13:43.931 nvmf hotplug test: fio failed as expected 00:13:43.931 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:43.931 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@85 -- # rm -f ./local-job0-0-verify.state 00:13:43.931 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@86 -- # rm -f ./local-job1-1-verify.state 00:13:43.931 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@87 -- # rm -f ./local-job2-2-verify.state 00:13:43.931 08:49:33 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@89 -- # trap - SIGINT SIGTERM EXIT 00:13:43.931 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- target/fio.sh@91 -- # nvmftestfini 00:13:43.931 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@514 -- # nvmfcleanup 00:13:43.931 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@121 -- # sync 00:13:43.931 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:43.931 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@124 -- # set +e 00:13:43.931 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:43.931 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:43.931 rmmod nvme_tcp 00:13:43.931 rmmod nvme_fabrics 00:13:44.192 rmmod nvme_keyring 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@128 -- # set -e 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@129 -- # return 0 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@515 -- # '[' -n 1752164 ']' 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@516 -- # killprocess 1752164 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@950 -- # '[' -z 1752164 ']' 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@954 -- # kill -0 1752164 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@955 -- # uname 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1752164 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1752164' 00:13:44.192 killing process with pid 1752164 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@969 -- # kill 1752164 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@974 -- # wait 1752164 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@297 -- # iptr 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@789 -- # iptables-save 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@789 -- # iptables-restore 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:44.192 08:49:34 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:46.735 00:13:46.735 real 0m28.980s 00:13:46.735 user 2m32.453s 00:13:46.735 sys 0m9.081s 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:13:46.735 ************************************ 00:13:46.735 END TEST nvmf_fio_target 00:13:46.735 ************************************ 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@35 -- # run_test nvmf_bdevio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:13:46.735 ************************************ 00:13:46.735 START TEST nvmf_bdevio 00:13:46.735 ************************************ 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:13:46.735 * Looking for test storage... 00:13:46.735 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1689 -- # lcov --version 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@336 -- # IFS=.-: 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@336 -- # read -ra ver1 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@337 -- # IFS=.-: 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@337 -- # read -ra ver2 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@338 -- # local 'op=<' 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@340 -- # ver1_l=2 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@341 -- # ver2_l=1 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@344 -- # case "$op" in 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@345 -- # : 1 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@365 -- # decimal 1 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@353 -- # local d=1 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@355 -- # echo 1 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@365 -- # ver1[v]=1 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@366 -- # decimal 2 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@353 -- # local d=2 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@355 -- # echo 2 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@366 -- # ver2[v]=2 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@368 -- # return 0 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:13:46.735 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:46.735 --rc genhtml_branch_coverage=1 00:13:46.735 --rc genhtml_function_coverage=1 00:13:46.735 --rc genhtml_legend=1 00:13:46.735 --rc geninfo_all_blocks=1 00:13:46.735 --rc geninfo_unexecuted_blocks=1 00:13:46.735 00:13:46.735 ' 00:13:46.735 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:13:46.735 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:46.736 --rc genhtml_branch_coverage=1 00:13:46.736 --rc genhtml_function_coverage=1 00:13:46.736 --rc genhtml_legend=1 00:13:46.736 --rc geninfo_all_blocks=1 00:13:46.736 --rc geninfo_unexecuted_blocks=1 00:13:46.736 00:13:46.736 ' 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:13:46.736 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:46.736 --rc genhtml_branch_coverage=1 00:13:46.736 --rc genhtml_function_coverage=1 00:13:46.736 --rc genhtml_legend=1 00:13:46.736 --rc geninfo_all_blocks=1 00:13:46.736 --rc geninfo_unexecuted_blocks=1 00:13:46.736 00:13:46.736 ' 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:13:46.736 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:46.736 --rc genhtml_branch_coverage=1 00:13:46.736 --rc genhtml_function_coverage=1 00:13:46.736 --rc genhtml_legend=1 00:13:46.736 --rc geninfo_all_blocks=1 00:13:46.736 --rc geninfo_unexecuted_blocks=1 00:13:46.736 00:13:46.736 ' 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@7 -- # uname -s 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@15 -- # shopt -s extglob 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@5 -- # export PATH 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@51 -- # : 0 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:46.736 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@14 -- # nvmftestinit 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@474 -- # prepare_net_devs 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@436 -- # local -g is_hw=no 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@438 -- # remove_spdk_ns 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@309 -- # xtrace_disable 00:13:46.736 08:49:36 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@315 -- # pci_devs=() 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@315 -- # local -a pci_devs 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@316 -- # pci_net_devs=() 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@317 -- # pci_drivers=() 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@317 -- # local -A pci_drivers 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@319 -- # net_devs=() 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@319 -- # local -ga net_devs 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@320 -- # e810=() 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@320 -- # local -ga e810 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@321 -- # x722=() 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@321 -- # local -ga x722 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@322 -- # mlx=() 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@322 -- # local -ga mlx 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:54.881 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:13:54.882 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:13:54.882 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:13:54.882 Found net devices under 0000:4b:00.0: cvl_0_0 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ up == up ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:13:54.882 Found net devices under 0000:4b:00.1: cvl_0_1 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@440 -- # is_hw=yes 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:13:54.882 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:54.882 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.598 ms 00:13:54.882 00:13:54.882 --- 10.0.0.2 ping statistics --- 00:13:54.882 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:54.882 rtt min/avg/max/mdev = 0.598/0.598/0.598/0.000 ms 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:54.882 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:54.882 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.177 ms 00:13:54.882 00:13:54.882 --- 10.0.0.1 ping statistics --- 00:13:54.882 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:54.882 rtt min/avg/max/mdev = 0.177/0.177/0.177/0.000 ms 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@448 -- # return 0 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@724 -- # xtrace_disable 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@507 -- # nvmfpid=1761190 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@508 -- # waitforlisten 1761190 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@831 -- # '[' -z 1761190 ']' 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@836 -- # local max_retries=100 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:54.882 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@840 -- # xtrace_disable 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:54.882 08:49:43 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x78 00:13:54.882 [2024-11-06 08:49:43.838118] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:13:54.882 [2024-11-06 08:49:43.838176] [ DPDK EAL parameters: nvmf -c 0x78 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:54.882 [2024-11-06 08:49:43.934740] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:54.882 [2024-11-06 08:49:43.987176] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:54.882 [2024-11-06 08:49:43.987249] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:54.882 [2024-11-06 08:49:43.987257] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:54.882 [2024-11-06 08:49:43.987264] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:54.882 [2024-11-06 08:49:43.987271] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:54.883 [2024-11-06 08:49:43.989319] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:13:54.883 [2024-11-06 08:49:43.989479] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:13:54.883 [2024-11-06 08:49:43.989638] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:13:54.883 [2024-11-06 08:49:43.989638] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@864 -- # return 0 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@730 -- # xtrace_disable 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:54.883 [2024-11-06 08:49:44.686473] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:54.883 Malloc0 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:54.883 [2024-11-06 08:49:44.761337] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@558 -- # config=() 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@558 -- # local subsystem config 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:13:54.883 { 00:13:54.883 "params": { 00:13:54.883 "name": "Nvme$subsystem", 00:13:54.883 "trtype": "$TEST_TRANSPORT", 00:13:54.883 "traddr": "$NVMF_FIRST_TARGET_IP", 00:13:54.883 "adrfam": "ipv4", 00:13:54.883 "trsvcid": "$NVMF_PORT", 00:13:54.883 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:13:54.883 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:13:54.883 "hdgst": ${hdgst:-false}, 00:13:54.883 "ddgst": ${ddgst:-false} 00:13:54.883 }, 00:13:54.883 "method": "bdev_nvme_attach_controller" 00:13:54.883 } 00:13:54.883 EOF 00:13:54.883 )") 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@580 -- # cat 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@582 -- # jq . 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@583 -- # IFS=, 00:13:54.883 08:49:44 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:13:54.883 "params": { 00:13:54.883 "name": "Nvme1", 00:13:54.883 "trtype": "tcp", 00:13:54.883 "traddr": "10.0.0.2", 00:13:54.883 "adrfam": "ipv4", 00:13:54.883 "trsvcid": "4420", 00:13:54.883 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:13:54.883 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:13:54.883 "hdgst": false, 00:13:54.883 "ddgst": false 00:13:54.883 }, 00:13:54.883 "method": "bdev_nvme_attach_controller" 00:13:54.883 }' 00:13:54.883 [2024-11-06 08:49:44.814362] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:13:54.883 [2024-11-06 08:49:44.814438] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1761459 ] 00:13:54.883 [2024-11-06 08:49:44.892807] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:13:54.883 [2024-11-06 08:49:44.937449] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:13:54.883 [2024-11-06 08:49:44.937578] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:13:54.883 [2024-11-06 08:49:44.937582] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:13:55.143 I/O targets: 00:13:55.143 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:13:55.143 00:13:55.143 00:13:55.143 CUnit - A unit testing framework for C - Version 2.1-3 00:13:55.143 http://cunit.sourceforge.net/ 00:13:55.143 00:13:55.143 00:13:55.143 Suite: bdevio tests on: Nvme1n1 00:13:55.143 Test: blockdev write read block ...passed 00:13:55.143 Test: blockdev write zeroes read block ...passed 00:13:55.143 Test: blockdev write zeroes read no split ...passed 00:13:55.143 Test: blockdev write zeroes read split ...passed 00:13:55.143 Test: blockdev write zeroes read split partial ...passed 00:13:55.143 Test: blockdev reset ...[2024-11-06 08:49:45.210717] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:13:55.143 [2024-11-06 08:49:45.210789] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x17f4970 (9): Bad file descriptor 00:13:55.404 [2024-11-06 08:49:45.361271] bdev_nvme.c:2236:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller successful. 00:13:55.404 passed 00:13:55.404 Test: blockdev write read 8 blocks ...passed 00:13:55.404 Test: blockdev write read size > 128k ...passed 00:13:55.404 Test: blockdev write read invalid size ...passed 00:13:55.404 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:13:55.404 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:13:55.404 Test: blockdev write read max offset ...passed 00:13:55.404 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:13:55.404 Test: blockdev writev readv 8 blocks ...passed 00:13:55.404 Test: blockdev writev readv 30 x 1block ...passed 00:13:55.667 Test: blockdev writev readv block ...passed 00:13:55.667 Test: blockdev writev readv size > 128k ...passed 00:13:55.667 Test: blockdev writev readv size > 128k in two iovs ...passed 00:13:55.667 Test: blockdev comparev and writev ...[2024-11-06 08:49:45.544105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:13:55.667 [2024-11-06 08:49:45.544130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:13:55.667 [2024-11-06 08:49:45.544142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:13:55.667 [2024-11-06 08:49:45.544148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:13:55.667 [2024-11-06 08:49:45.544618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:13:55.667 [2024-11-06 08:49:45.544627] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:13:55.667 [2024-11-06 08:49:45.544637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:13:55.667 [2024-11-06 08:49:45.544643] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:13:55.667 [2024-11-06 08:49:45.545087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:13:55.667 [2024-11-06 08:49:45.545096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:13:55.667 [2024-11-06 08:49:45.545105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:13:55.667 [2024-11-06 08:49:45.545111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:13:55.667 [2024-11-06 08:49:45.545538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:13:55.667 [2024-11-06 08:49:45.545546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:13:55.667 [2024-11-06 08:49:45.545560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:13:55.667 [2024-11-06 08:49:45.545566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:13:55.667 passed 00:13:55.667 Test: blockdev nvme passthru rw ...passed 00:13:55.667 Test: blockdev nvme passthru vendor specific ...[2024-11-06 08:49:45.630347] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:13:55.667 [2024-11-06 08:49:45.630357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:13:55.667 [2024-11-06 08:49:45.630689] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:13:55.667 [2024-11-06 08:49:45.630696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:13:55.667 [2024-11-06 08:49:45.631056] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:13:55.667 [2024-11-06 08:49:45.631065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:13:55.667 [2024-11-06 08:49:45.631404] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:13:55.667 [2024-11-06 08:49:45.631412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:13:55.667 passed 00:13:55.667 Test: blockdev nvme admin passthru ...passed 00:13:55.667 Test: blockdev copy ...passed 00:13:55.667 00:13:55.667 Run Summary: Type Total Ran Passed Failed Inactive 00:13:55.667 suites 1 1 n/a 0 0 00:13:55.667 tests 23 23 23 0 0 00:13:55.667 asserts 152 152 152 0 n/a 00:13:55.667 00:13:55.667 Elapsed time = 1.214 seconds 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- target/bdevio.sh@30 -- # nvmftestfini 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@514 -- # nvmfcleanup 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@121 -- # sync 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@124 -- # set +e 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@125 -- # for i in {1..20} 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:13:55.929 rmmod nvme_tcp 00:13:55.929 rmmod nvme_fabrics 00:13:55.929 rmmod nvme_keyring 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@128 -- # set -e 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@129 -- # return 0 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@515 -- # '[' -n 1761190 ']' 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@516 -- # killprocess 1761190 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@950 -- # '[' -z 1761190 ']' 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@954 -- # kill -0 1761190 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@955 -- # uname 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1761190 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@956 -- # process_name=reactor_3 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@960 -- # '[' reactor_3 = sudo ']' 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1761190' 00:13:55.929 killing process with pid 1761190 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@969 -- # kill 1761190 00:13:55.929 08:49:45 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@974 -- # wait 1761190 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@297 -- # iptr 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@789 -- # iptables-save 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@789 -- # iptables-restore 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@302 -- # remove_spdk_ns 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:56.190 08:49:46 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:58.106 08:49:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:13:58.106 00:13:58.106 real 0m11.764s 00:13:58.106 user 0m12.555s 00:13:58.106 sys 0m5.903s 00:13:58.106 08:49:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:58.106 08:49:48 nvmf_tcp.nvmf_target_core.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:13:58.106 ************************************ 00:13:58.106 END TEST nvmf_bdevio 00:13:58.106 ************************************ 00:13:58.367 08:49:48 nvmf_tcp.nvmf_target_core -- nvmf/nvmf_target_core.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:13:58.367 00:13:58.367 real 4m59.211s 00:13:58.367 user 11m39.343s 00:13:58.367 sys 1m46.938s 00:13:58.367 08:49:48 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@1126 -- # xtrace_disable 00:13:58.367 08:49:48 nvmf_tcp.nvmf_target_core -- common/autotest_common.sh@10 -- # set +x 00:13:58.367 ************************************ 00:13:58.367 END TEST nvmf_target_core 00:13:58.367 ************************************ 00:13:58.367 08:49:48 nvmf_tcp -- nvmf/nvmf.sh@15 -- # run_test nvmf_target_extra /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_extra.sh --transport=tcp 00:13:58.367 08:49:48 nvmf_tcp -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:13:58.367 08:49:48 nvmf_tcp -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:58.367 08:49:48 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:13:58.367 ************************************ 00:13:58.367 START TEST nvmf_target_extra 00:13:58.367 ************************************ 00:13:58.367 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_extra.sh --transport=tcp 00:13:58.367 * Looking for test storage... 00:13:58.367 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:13:58.367 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:13:58.367 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1689 -- # lcov --version 00:13:58.367 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@336 -- # IFS=.-: 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@336 -- # read -ra ver1 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@337 -- # IFS=.-: 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@337 -- # read -ra ver2 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@338 -- # local 'op=<' 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@340 -- # ver1_l=2 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@341 -- # ver2_l=1 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@344 -- # case "$op" in 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@345 -- # : 1 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@365 -- # decimal 1 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@353 -- # local d=1 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@355 -- # echo 1 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@365 -- # ver1[v]=1 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@366 -- # decimal 2 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@353 -- # local d=2 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@355 -- # echo 2 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@366 -- # ver2[v]=2 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@368 -- # return 0 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:13:58.630 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:58.630 --rc genhtml_branch_coverage=1 00:13:58.630 --rc genhtml_function_coverage=1 00:13:58.630 --rc genhtml_legend=1 00:13:58.630 --rc geninfo_all_blocks=1 00:13:58.630 --rc geninfo_unexecuted_blocks=1 00:13:58.630 00:13:58.630 ' 00:13:58.630 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:13:58.630 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:58.630 --rc genhtml_branch_coverage=1 00:13:58.630 --rc genhtml_function_coverage=1 00:13:58.631 --rc genhtml_legend=1 00:13:58.631 --rc geninfo_all_blocks=1 00:13:58.631 --rc geninfo_unexecuted_blocks=1 00:13:58.631 00:13:58.631 ' 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:13:58.631 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:58.631 --rc genhtml_branch_coverage=1 00:13:58.631 --rc genhtml_function_coverage=1 00:13:58.631 --rc genhtml_legend=1 00:13:58.631 --rc geninfo_all_blocks=1 00:13:58.631 --rc geninfo_unexecuted_blocks=1 00:13:58.631 00:13:58.631 ' 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:13:58.631 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:58.631 --rc genhtml_branch_coverage=1 00:13:58.631 --rc genhtml_function_coverage=1 00:13:58.631 --rc genhtml_legend=1 00:13:58.631 --rc geninfo_all_blocks=1 00:13:58.631 --rc geninfo_unexecuted_blocks=1 00:13:58.631 00:13:58.631 ' 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@7 -- # uname -s 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@15 -- # shopt -s extglob 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- paths/export.sh@5 -- # export PATH 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@51 -- # : 0 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:58.631 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@11 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@13 -- # TEST_ARGS=("$@") 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@15 -- # [[ 0 -eq 0 ]] 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@16 -- # run_test nvmf_example /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:13:58.631 ************************************ 00:13:58.631 START TEST nvmf_example 00:13:58.631 ************************************ 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:13:58.631 * Looking for test storage... 00:13:58.631 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1689 -- # lcov --version 00:13:58.631 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@333 -- # local ver1 ver1_l 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@334 -- # local ver2 ver2_l 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@336 -- # IFS=.-: 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@336 -- # read -ra ver1 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@337 -- # IFS=.-: 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@337 -- # read -ra ver2 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@338 -- # local 'op=<' 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@340 -- # ver1_l=2 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@341 -- # ver2_l=1 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@344 -- # case "$op" in 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@345 -- # : 1 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@364 -- # (( v = 0 )) 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@365 -- # decimal 1 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@353 -- # local d=1 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@355 -- # echo 1 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@365 -- # ver1[v]=1 00:13:58.894 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@366 -- # decimal 2 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@353 -- # local d=2 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@355 -- # echo 2 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@366 -- # ver2[v]=2 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@368 -- # return 0 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:13:58.895 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:58.895 --rc genhtml_branch_coverage=1 00:13:58.895 --rc genhtml_function_coverage=1 00:13:58.895 --rc genhtml_legend=1 00:13:58.895 --rc geninfo_all_blocks=1 00:13:58.895 --rc geninfo_unexecuted_blocks=1 00:13:58.895 00:13:58.895 ' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:13:58.895 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:58.895 --rc genhtml_branch_coverage=1 00:13:58.895 --rc genhtml_function_coverage=1 00:13:58.895 --rc genhtml_legend=1 00:13:58.895 --rc geninfo_all_blocks=1 00:13:58.895 --rc geninfo_unexecuted_blocks=1 00:13:58.895 00:13:58.895 ' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:13:58.895 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:58.895 --rc genhtml_branch_coverage=1 00:13:58.895 --rc genhtml_function_coverage=1 00:13:58.895 --rc genhtml_legend=1 00:13:58.895 --rc geninfo_all_blocks=1 00:13:58.895 --rc geninfo_unexecuted_blocks=1 00:13:58.895 00:13:58.895 ' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:13:58.895 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:13:58.895 --rc genhtml_branch_coverage=1 00:13:58.895 --rc genhtml_function_coverage=1 00:13:58.895 --rc genhtml_legend=1 00:13:58.895 --rc geninfo_all_blocks=1 00:13:58.895 --rc geninfo_unexecuted_blocks=1 00:13:58.895 00:13:58.895 ' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@7 -- # uname -s 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@15 -- # shopt -s extglob 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@5 -- # export PATH 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@51 -- # : 0 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:13:58.895 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@55 -- # have_pci_nics=0 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@11 -- # NVMF_EXAMPLE=("$SPDK_EXAMPLE_DIR/nvmf") 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@13 -- # MALLOC_BDEV_SIZE=64 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@24 -- # build_nvmf_example_args 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@17 -- # '[' 0 -eq 1 ']' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@20 -- # NVMF_EXAMPLE+=(-i "$NVMF_APP_SHM_ID" -g 10000) 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@21 -- # NVMF_EXAMPLE+=("${NO_HUGE[@]}") 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@40 -- # timing_enter nvmf_example_test 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@724 -- # xtrace_disable 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@41 -- # nvmftestinit 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@474 -- # prepare_net_devs 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@436 -- # local -g is_hw=no 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@438 -- # remove_spdk_ns 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@309 -- # xtrace_disable 00:13:58.895 08:49:48 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@315 -- # pci_devs=() 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@315 -- # local -a pci_devs 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@316 -- # pci_net_devs=() 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@317 -- # pci_drivers=() 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@317 -- # local -A pci_drivers 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@319 -- # net_devs=() 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@319 -- # local -ga net_devs 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@320 -- # e810=() 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@320 -- # local -ga e810 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@321 -- # x722=() 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@321 -- # local -ga x722 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@322 -- # mlx=() 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@322 -- # local -ga mlx 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:14:07.040 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:14:07.040 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@416 -- # [[ up == up ]] 00:14:07.040 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:14:07.041 Found net devices under 0000:4b:00.0: cvl_0_0 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@416 -- # [[ up == up ]] 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:14:07.041 Found net devices under 0000:4b:00.1: cvl_0_1 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@440 -- # is_hw=yes 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:14:07.041 08:49:55 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:14:07.041 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:07.041 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.579 ms 00:14:07.041 00:14:07.041 --- 10.0.0.2 ping statistics --- 00:14:07.041 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:07.041 rtt min/avg/max/mdev = 0.579/0.579/0.579/0.000 ms 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:07.041 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:07.041 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.297 ms 00:14:07.041 00:14:07.041 --- 10.0.0.1 ping statistics --- 00:14:07.041 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:07.041 rtt min/avg/max/mdev = 0.297/0.297/0.297/0.000 ms 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@448 -- # return 0 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@42 -- # nvmfexamplestart '-m 0xF' 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@27 -- # timing_enter start_nvmf_example 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@724 -- # xtrace_disable 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@29 -- # '[' tcp == tcp ']' 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@30 -- # NVMF_EXAMPLE=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_EXAMPLE[@]}") 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@34 -- # nvmfpid=1766085 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@35 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/nvmf -i 0 -g 10000 -m 0xF 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@36 -- # waitforlisten 1766085 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@831 -- # '[' -z 1766085 ']' 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:07.041 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@864 -- # return 0 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@37 -- # timing_exit start_nvmf_example 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@730 -- # xtrace_disable 00:14:07.041 08:49:56 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@47 -- # rpc_cmd bdev_malloc_create 64 512 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@47 -- # malloc_bdevs='Malloc0 ' 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@49 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@52 -- # for malloc_bdev in $malloc_bdevs 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@57 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:07.041 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@59 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:14:07.042 08:49:57 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:14:19.292 Initializing NVMe Controllers 00:14:19.292 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:14:19.292 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:14:19.292 Initialization complete. Launching workers. 00:14:19.292 ======================================================== 00:14:19.292 Latency(us) 00:14:19.292 Device Information : IOPS MiB/s Average min max 00:14:19.292 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 18270.26 71.37 3502.47 708.43 17300.74 00:14:19.292 ======================================================== 00:14:19.292 Total : 18270.26 71.37 3502.47 708.43 17300.74 00:14:19.292 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@65 -- # trap - SIGINT SIGTERM EXIT 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@66 -- # nvmftestfini 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@514 -- # nvmfcleanup 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@121 -- # sync 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@124 -- # set +e 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@125 -- # for i in {1..20} 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:14:19.292 rmmod nvme_tcp 00:14:19.292 rmmod nvme_fabrics 00:14:19.292 rmmod nvme_keyring 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@128 -- # set -e 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@129 -- # return 0 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@515 -- # '[' -n 1766085 ']' 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@516 -- # killprocess 1766085 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@950 -- # '[' -z 1766085 ']' 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@954 -- # kill -0 1766085 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@955 -- # uname 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1766085 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@956 -- # process_name=nvmf 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@960 -- # '[' nvmf = sudo ']' 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1766085' 00:14:19.292 killing process with pid 1766085 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@969 -- # kill 1766085 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@974 -- # wait 1766085 00:14:19.292 nvmf threads initialize successfully 00:14:19.292 bdev subsystem init successfully 00:14:19.292 created a nvmf target service 00:14:19.292 create targets's poll groups done 00:14:19.292 all subsystems of target started 00:14:19.292 nvmf target is running 00:14:19.292 all subsystems of target stopped 00:14:19.292 destroy targets's poll groups done 00:14:19.292 destroyed the nvmf target service 00:14:19.292 bdev subsystem finish successfully 00:14:19.292 nvmf threads destroy successfully 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@297 -- # iptr 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@789 -- # iptables-save 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@789 -- # iptables-restore 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@302 -- # remove_spdk_ns 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:19.292 08:50:07 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra.nvmf_example -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra.nvmf_example -- target/nvmf_example.sh@67 -- # timing_exit nvmf_example_test 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@730 -- # xtrace_disable 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:19.862 00:14:19.862 real 0m21.265s 00:14:19.862 user 0m47.101s 00:14:19.862 sys 0m6.629s 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:14:19.862 ************************************ 00:14:19.862 END TEST nvmf_example 00:14:19.862 ************************************ 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@17 -- # run_test nvmf_filesystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:14:19.862 ************************************ 00:14:19.862 START TEST nvmf_filesystem 00:14:19.862 ************************************ 00:14:19.862 08:50:09 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:14:20.125 * Looking for test storage... 00:14:20.125 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1689 -- # lcov --version 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # IFS=.-: 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # read -ra ver1 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # IFS=.-: 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # read -ra ver2 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@338 -- # local 'op=<' 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@340 -- # ver1_l=2 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@341 -- # ver2_l=1 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@344 -- # case "$op" in 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@345 -- # : 1 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # decimal 1 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=1 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 1 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # ver1[v]=1 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # decimal 2 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=2 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 2 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # ver2[v]=2 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # return 0 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:14:20.125 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:20.125 --rc genhtml_branch_coverage=1 00:14:20.125 --rc genhtml_function_coverage=1 00:14:20.125 --rc genhtml_legend=1 00:14:20.125 --rc geninfo_all_blocks=1 00:14:20.125 --rc geninfo_unexecuted_blocks=1 00:14:20.125 00:14:20.125 ' 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:14:20.125 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:20.125 --rc genhtml_branch_coverage=1 00:14:20.125 --rc genhtml_function_coverage=1 00:14:20.125 --rc genhtml_legend=1 00:14:20.125 --rc geninfo_all_blocks=1 00:14:20.125 --rc geninfo_unexecuted_blocks=1 00:14:20.125 00:14:20.125 ' 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:14:20.125 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:20.125 --rc genhtml_branch_coverage=1 00:14:20.125 --rc genhtml_function_coverage=1 00:14:20.125 --rc genhtml_legend=1 00:14:20.125 --rc geninfo_all_blocks=1 00:14:20.125 --rc geninfo_unexecuted_blocks=1 00:14:20.125 00:14:20.125 ' 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:14:20.125 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:20.125 --rc genhtml_branch_coverage=1 00:14:20.125 --rc genhtml_function_coverage=1 00:14:20.125 --rc genhtml_legend=1 00:14:20.125 --rc geninfo_all_blocks=1 00:14:20.125 --rc geninfo_unexecuted_blocks=1 00:14:20.125 00:14:20.125 ' 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@7 -- # rpc_py=rpc_cmd 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@34 -- # set -e 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@35 -- # shopt -s nullglob 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@36 -- # shopt -s extglob 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@37 -- # shopt -s inherit_errexit 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@39 -- # '[' -z /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output ']' 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@44 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh ]] 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@1 -- # CONFIG_WPDK_DIR= 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@2 -- # CONFIG_ASAN=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@3 -- # CONFIG_VBDEV_COMPRESS=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@4 -- # CONFIG_HAVE_EXECINFO_H=y 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@5 -- # CONFIG_USDT=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@6 -- # CONFIG_CUSTOMOCF=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@7 -- # CONFIG_PREFIX=/usr/local 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@8 -- # CONFIG_RBD=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@9 -- # CONFIG_LIBDIR= 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@10 -- # CONFIG_IDXD=y 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@11 -- # CONFIG_NVME_CUSE=y 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@12 -- # CONFIG_SMA=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@13 -- # CONFIG_VTUNE=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@14 -- # CONFIG_TSAN=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@15 -- # CONFIG_RDMA_SEND_WITH_INVAL=y 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@16 -- # CONFIG_VFIO_USER_DIR= 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@17 -- # CONFIG_MAX_NUMA_NODES=1 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@18 -- # CONFIG_PGO_CAPTURE=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@19 -- # CONFIG_HAVE_UUID_GENERATE_SHA1=y 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@20 -- # CONFIG_ENV=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@21 -- # CONFIG_LTO=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@22 -- # CONFIG_ISCSI_INITIATOR=y 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@23 -- # CONFIG_CET=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@24 -- # CONFIG_VBDEV_COMPRESS_MLX5=n 00:14:20.125 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@25 -- # CONFIG_OCF_PATH= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@26 -- # CONFIG_RDMA_SET_TOS=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@27 -- # CONFIG_AIO_FSDEV=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@28 -- # CONFIG_HAVE_ARC4RANDOM=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@29 -- # CONFIG_HAVE_LIBARCHIVE=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@30 -- # CONFIG_UBLK=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@31 -- # CONFIG_ISAL_CRYPTO=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@32 -- # CONFIG_OPENSSL_PATH= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@33 -- # CONFIG_OCF=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@34 -- # CONFIG_FUSE=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@35 -- # CONFIG_VTUNE_DIR= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@36 -- # CONFIG_FUZZER_LIB= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@37 -- # CONFIG_FUZZER=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@38 -- # CONFIG_FSDEV=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@39 -- # CONFIG_DPDK_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@40 -- # CONFIG_CRYPTO=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@41 -- # CONFIG_PGO_USE=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@42 -- # CONFIG_VHOST=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@43 -- # CONFIG_DAOS=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@44 -- # CONFIG_DPDK_INC_DIR= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@45 -- # CONFIG_DAOS_DIR= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@46 -- # CONFIG_UNIT_TESTS=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@47 -- # CONFIG_RDMA_SET_ACK_TIMEOUT=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@48 -- # CONFIG_VIRTIO=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@49 -- # CONFIG_DPDK_UADK=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@50 -- # CONFIG_COVERAGE=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@51 -- # CONFIG_RDMA=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@52 -- # CONFIG_HAVE_STRUCT_STAT_ST_ATIM=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@53 -- # CONFIG_HAVE_LZ4=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@54 -- # CONFIG_FIO_SOURCE_DIR=/usr/src/fio 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@55 -- # CONFIG_URING_PATH= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@56 -- # CONFIG_XNVME=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@57 -- # CONFIG_VFIO_USER=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@58 -- # CONFIG_ARCH=native 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@59 -- # CONFIG_HAVE_EVP_MAC=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@60 -- # CONFIG_URING_ZNS=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@61 -- # CONFIG_WERROR=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@62 -- # CONFIG_HAVE_LIBBSD=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@63 -- # CONFIG_UBSAN=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@64 -- # CONFIG_HAVE_STRUCT_STAT_ST_ATIMESPEC=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@65 -- # CONFIG_IPSEC_MB_DIR= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@66 -- # CONFIG_GOLANG=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@67 -- # CONFIG_ISAL=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@68 -- # CONFIG_IDXD_KERNEL=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@69 -- # CONFIG_DPDK_LIB_DIR= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@70 -- # CONFIG_RDMA_PROV=verbs 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@71 -- # CONFIG_APPS=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@72 -- # CONFIG_SHARED=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@73 -- # CONFIG_HAVE_KEYUTILS=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@74 -- # CONFIG_FC_PATH= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@75 -- # CONFIG_DPDK_PKG_CONFIG=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@76 -- # CONFIG_FC=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@77 -- # CONFIG_AVAHI=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@78 -- # CONFIG_FIO_PLUGIN=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@79 -- # CONFIG_RAID5F=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@80 -- # CONFIG_EXAMPLES=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@81 -- # CONFIG_TESTS=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@82 -- # CONFIG_CRYPTO_MLX5=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@83 -- # CONFIG_MAX_LCORES=128 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@84 -- # CONFIG_IPSEC_MB=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@85 -- # CONFIG_PGO_DIR= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@86 -- # CONFIG_DEBUG=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@87 -- # CONFIG_DPDK_COMPRESSDEV=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@88 -- # CONFIG_CROSS_PREFIX= 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@89 -- # CONFIG_COPY_FILE_RANGE=y 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/build_config.sh@90 -- # CONFIG_URING=n 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@54 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@8 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@9 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@10 -- # _app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@11 -- # _test_app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@12 -- # _examples_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@14 -- # VHOST_FUZZ_APP=("$_test_app_dir/fuzz/vhost_fuzz/vhost_fuzz") 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@15 -- # ISCSI_APP=("$_app_dir/iscsi_tgt") 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@16 -- # NVMF_APP=("$_app_dir/nvmf_tgt") 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@17 -- # VHOST_APP=("$_app_dir/vhost") 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@18 -- # DD_APP=("$_app_dir/spdk_dd") 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@19 -- # SPDK_APP=("$_app_dir/spdk_tgt") 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@22 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/config.h ]] 00:14:20.126 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@23 -- # [[ #ifndef SPDK_CONFIG_H 00:14:20.126 #define SPDK_CONFIG_H 00:14:20.126 #define SPDK_CONFIG_AIO_FSDEV 1 00:14:20.126 #define SPDK_CONFIG_APPS 1 00:14:20.126 #define SPDK_CONFIG_ARCH native 00:14:20.126 #undef SPDK_CONFIG_ASAN 00:14:20.126 #undef SPDK_CONFIG_AVAHI 00:14:20.126 #undef SPDK_CONFIG_CET 00:14:20.126 #define SPDK_CONFIG_COPY_FILE_RANGE 1 00:14:20.126 #define SPDK_CONFIG_COVERAGE 1 00:14:20.126 #define SPDK_CONFIG_CROSS_PREFIX 00:14:20.126 #undef SPDK_CONFIG_CRYPTO 00:14:20.126 #undef SPDK_CONFIG_CRYPTO_MLX5 00:14:20.126 #undef SPDK_CONFIG_CUSTOMOCF 00:14:20.127 #undef SPDK_CONFIG_DAOS 00:14:20.127 #define SPDK_CONFIG_DAOS_DIR 00:14:20.127 #define SPDK_CONFIG_DEBUG 1 00:14:20.127 #undef SPDK_CONFIG_DPDK_COMPRESSDEV 00:14:20.127 #define SPDK_CONFIG_DPDK_DIR /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:14:20.127 #define SPDK_CONFIG_DPDK_INC_DIR 00:14:20.127 #define SPDK_CONFIG_DPDK_LIB_DIR 00:14:20.127 #undef SPDK_CONFIG_DPDK_PKG_CONFIG 00:14:20.127 #undef SPDK_CONFIG_DPDK_UADK 00:14:20.127 #define SPDK_CONFIG_ENV /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:14:20.127 #define SPDK_CONFIG_EXAMPLES 1 00:14:20.127 #undef SPDK_CONFIG_FC 00:14:20.127 #define SPDK_CONFIG_FC_PATH 00:14:20.127 #define SPDK_CONFIG_FIO_PLUGIN 1 00:14:20.127 #define SPDK_CONFIG_FIO_SOURCE_DIR /usr/src/fio 00:14:20.127 #define SPDK_CONFIG_FSDEV 1 00:14:20.127 #undef SPDK_CONFIG_FUSE 00:14:20.127 #undef SPDK_CONFIG_FUZZER 00:14:20.127 #define SPDK_CONFIG_FUZZER_LIB 00:14:20.127 #undef SPDK_CONFIG_GOLANG 00:14:20.127 #define SPDK_CONFIG_HAVE_ARC4RANDOM 1 00:14:20.127 #define SPDK_CONFIG_HAVE_EVP_MAC 1 00:14:20.127 #define SPDK_CONFIG_HAVE_EXECINFO_H 1 00:14:20.127 #define SPDK_CONFIG_HAVE_KEYUTILS 1 00:14:20.127 #undef SPDK_CONFIG_HAVE_LIBARCHIVE 00:14:20.127 #undef SPDK_CONFIG_HAVE_LIBBSD 00:14:20.127 #undef SPDK_CONFIG_HAVE_LZ4 00:14:20.127 #define SPDK_CONFIG_HAVE_STRUCT_STAT_ST_ATIM 1 00:14:20.127 #undef SPDK_CONFIG_HAVE_STRUCT_STAT_ST_ATIMESPEC 00:14:20.127 #define SPDK_CONFIG_HAVE_UUID_GENERATE_SHA1 1 00:14:20.127 #define SPDK_CONFIG_IDXD 1 00:14:20.127 #define SPDK_CONFIG_IDXD_KERNEL 1 00:14:20.127 #undef SPDK_CONFIG_IPSEC_MB 00:14:20.127 #define SPDK_CONFIG_IPSEC_MB_DIR 00:14:20.127 #define SPDK_CONFIG_ISAL 1 00:14:20.127 #define SPDK_CONFIG_ISAL_CRYPTO 1 00:14:20.127 #define SPDK_CONFIG_ISCSI_INITIATOR 1 00:14:20.127 #define SPDK_CONFIG_LIBDIR 00:14:20.127 #undef SPDK_CONFIG_LTO 00:14:20.127 #define SPDK_CONFIG_MAX_LCORES 128 00:14:20.127 #define SPDK_CONFIG_MAX_NUMA_NODES 1 00:14:20.127 #define SPDK_CONFIG_NVME_CUSE 1 00:14:20.127 #undef SPDK_CONFIG_OCF 00:14:20.127 #define SPDK_CONFIG_OCF_PATH 00:14:20.127 #define SPDK_CONFIG_OPENSSL_PATH 00:14:20.127 #undef SPDK_CONFIG_PGO_CAPTURE 00:14:20.127 #define SPDK_CONFIG_PGO_DIR 00:14:20.127 #undef SPDK_CONFIG_PGO_USE 00:14:20.127 #define SPDK_CONFIG_PREFIX /usr/local 00:14:20.127 #undef SPDK_CONFIG_RAID5F 00:14:20.127 #undef SPDK_CONFIG_RBD 00:14:20.127 #define SPDK_CONFIG_RDMA 1 00:14:20.127 #define SPDK_CONFIG_RDMA_PROV verbs 00:14:20.127 #define SPDK_CONFIG_RDMA_SEND_WITH_INVAL 1 00:14:20.127 #define SPDK_CONFIG_RDMA_SET_ACK_TIMEOUT 1 00:14:20.127 #define SPDK_CONFIG_RDMA_SET_TOS 1 00:14:20.127 #define SPDK_CONFIG_SHARED 1 00:14:20.127 #undef SPDK_CONFIG_SMA 00:14:20.127 #define SPDK_CONFIG_TESTS 1 00:14:20.127 #undef SPDK_CONFIG_TSAN 00:14:20.127 #define SPDK_CONFIG_UBLK 1 00:14:20.127 #define SPDK_CONFIG_UBSAN 1 00:14:20.127 #undef SPDK_CONFIG_UNIT_TESTS 00:14:20.127 #undef SPDK_CONFIG_URING 00:14:20.127 #define SPDK_CONFIG_URING_PATH 00:14:20.127 #undef SPDK_CONFIG_URING_ZNS 00:14:20.127 #undef SPDK_CONFIG_USDT 00:14:20.127 #undef SPDK_CONFIG_VBDEV_COMPRESS 00:14:20.127 #undef SPDK_CONFIG_VBDEV_COMPRESS_MLX5 00:14:20.127 #define SPDK_CONFIG_VFIO_USER 1 00:14:20.127 #define SPDK_CONFIG_VFIO_USER_DIR 00:14:20.127 #define SPDK_CONFIG_VHOST 1 00:14:20.127 #define SPDK_CONFIG_VIRTIO 1 00:14:20.127 #undef SPDK_CONFIG_VTUNE 00:14:20.127 #define SPDK_CONFIG_VTUNE_DIR 00:14:20.127 #define SPDK_CONFIG_WERROR 1 00:14:20.127 #define SPDK_CONFIG_WPDK_DIR 00:14:20.127 #undef SPDK_CONFIG_XNVME 00:14:20.127 #endif /* SPDK_CONFIG_H */ == *\#\d\e\f\i\n\e\ \S\P\D\K\_\C\O\N\F\I\G\_\D\E\B\U\G* ]] 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/applications.sh@24 -- # (( SPDK_AUTOTEST_DEBUG_APPS )) 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@55 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@15 -- # shopt -s extglob 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@5 -- # export PATH 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@56 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@6 -- # _pmdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@7 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/../../../ 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@7 -- # _pmrootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@64 -- # TEST_TAG=N/A 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@65 -- # TEST_TAG_FILE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.run_test_name 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@67 -- # PM_OUTPUTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@68 -- # uname -s 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@68 -- # PM_OS=Linux 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@70 -- # MONITOR_RESOURCES_SUDO=() 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@70 -- # declare -A MONITOR_RESOURCES_SUDO 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@71 -- # MONITOR_RESOURCES_SUDO["collect-bmc-pm"]=1 00:14:20.127 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@72 -- # MONITOR_RESOURCES_SUDO["collect-cpu-load"]=0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@73 -- # MONITOR_RESOURCES_SUDO["collect-cpu-temp"]=0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@74 -- # MONITOR_RESOURCES_SUDO["collect-vmstat"]=0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@76 -- # SUDO[0]= 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@76 -- # SUDO[1]='sudo -E' 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@78 -- # MONITOR_RESOURCES=(collect-cpu-load collect-vmstat) 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@79 -- # [[ Linux == FreeBSD ]] 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ Linux == Linux ]] 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ ............................... != QEMU ]] 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@81 -- # [[ ! -e /.dockerenv ]] 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@84 -- # MONITOR_RESOURCES+=(collect-cpu-temp) 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@85 -- # MONITOR_RESOURCES+=(collect-bmc-pm) 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- pm/common@88 -- # [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power ]] 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@58 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@59 -- # export RUN_NIGHTLY 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@62 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@63 -- # export SPDK_AUTOTEST_DEBUG_APPS 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@64 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@65 -- # export SPDK_RUN_VALGRIND 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@66 -- # : 1 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@67 -- # export SPDK_RUN_FUNCTIONAL_TEST 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@68 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@69 -- # export SPDK_TEST_UNITTEST 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@70 -- # : 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@71 -- # export SPDK_TEST_AUTOBUILD 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@72 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@73 -- # export SPDK_TEST_RELEASE_BUILD 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@74 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@75 -- # export SPDK_TEST_ISAL 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@76 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@77 -- # export SPDK_TEST_ISCSI 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@78 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@79 -- # export SPDK_TEST_ISCSI_INITIATOR 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@80 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@81 -- # export SPDK_TEST_NVME 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@82 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@83 -- # export SPDK_TEST_NVME_PMR 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@84 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@85 -- # export SPDK_TEST_NVME_BP 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@86 -- # : 1 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@87 -- # export SPDK_TEST_NVME_CLI 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@88 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@89 -- # export SPDK_TEST_NVME_CUSE 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@90 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@91 -- # export SPDK_TEST_NVME_FDP 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@92 -- # : 1 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@93 -- # export SPDK_TEST_NVMF 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@94 -- # : 1 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@95 -- # export SPDK_TEST_VFIOUSER 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@96 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@97 -- # export SPDK_TEST_VFIOUSER_QEMU 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@98 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@99 -- # export SPDK_TEST_FUZZER 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@100 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@101 -- # export SPDK_TEST_FUZZER_SHORT 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@102 -- # : tcp 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@103 -- # export SPDK_TEST_NVMF_TRANSPORT 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@104 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@105 -- # export SPDK_TEST_RBD 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@106 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@107 -- # export SPDK_TEST_VHOST 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@108 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@109 -- # export SPDK_TEST_BLOCKDEV 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@110 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@111 -- # export SPDK_TEST_RAID 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@112 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@113 -- # export SPDK_TEST_IOAT 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@114 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@115 -- # export SPDK_TEST_BLOBFS 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@116 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@117 -- # export SPDK_TEST_VHOST_INIT 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@118 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@119 -- # export SPDK_TEST_LVOL 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@120 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@121 -- # export SPDK_TEST_VBDEV_COMPRESS 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@122 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@123 -- # export SPDK_RUN_ASAN 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@124 -- # : 1 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@125 -- # export SPDK_RUN_UBSAN 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@126 -- # : 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@127 -- # export SPDK_RUN_EXTERNAL_DPDK 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@128 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@129 -- # export SPDK_RUN_NON_ROOT 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@130 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@131 -- # export SPDK_TEST_CRYPTO 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@132 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@133 -- # export SPDK_TEST_FTL 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@134 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@135 -- # export SPDK_TEST_OCF 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@136 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@137 -- # export SPDK_TEST_VMD 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@138 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@139 -- # export SPDK_TEST_OPAL 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@140 -- # : 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@141 -- # export SPDK_TEST_NATIVE_DPDK 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@142 -- # : true 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@143 -- # export SPDK_AUTOTEST_X 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@144 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@145 -- # export SPDK_TEST_URING 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@146 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@147 -- # export SPDK_TEST_USDT 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@148 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@149 -- # export SPDK_TEST_USE_IGB_UIO 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@150 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@151 -- # export SPDK_TEST_SCHEDULER 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@152 -- # : 0 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@153 -- # export SPDK_TEST_SCANBUILD 00:14:20.128 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@154 -- # : e810 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@155 -- # export SPDK_TEST_NVMF_NICS 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@156 -- # : 0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@157 -- # export SPDK_TEST_SMA 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@158 -- # : 0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@159 -- # export SPDK_TEST_DAOS 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@160 -- # : 0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@161 -- # export SPDK_TEST_XNVME 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@162 -- # : 0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@163 -- # export SPDK_TEST_ACCEL 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@164 -- # : 0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@165 -- # export SPDK_TEST_ACCEL_DSA 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@166 -- # : 0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@167 -- # export SPDK_TEST_ACCEL_IAA 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@169 -- # : 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@170 -- # export SPDK_TEST_FUZZER_TARGET 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@171 -- # : 0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@172 -- # export SPDK_TEST_NVMF_MDNS 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@173 -- # : 0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@174 -- # export SPDK_JSONRPC_GO_CLIENT 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@175 -- # : 0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@176 -- # export SPDK_TEST_SETUP 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@179 -- # export SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@179 -- # SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@180 -- # export DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@180 -- # DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@181 -- # export VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@181 -- # VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@182 -- # export LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@182 -- # LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@185 -- # export PCI_BLOCK_SYNC_ON_RESET=yes 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@185 -- # PCI_BLOCK_SYNC_ON_RESET=yes 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@189 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@189 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@193 -- # export PYTHONDONTWRITEBYTECODE=1 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@193 -- # PYTHONDONTWRITEBYTECODE=1 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@197 -- # export ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@197 -- # ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@198 -- # export UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@198 -- # UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@202 -- # asan_suppression_file=/var/tmp/asan_suppression_file 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@203 -- # rm -rf /var/tmp/asan_suppression_file 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@204 -- # cat 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@240 -- # echo leak:libfuse3.so 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@242 -- # export LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@242 -- # LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@244 -- # export DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@244 -- # DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@246 -- # '[' -z /var/spdk/dependencies ']' 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@249 -- # export DEPENDENCY_DIR 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@253 -- # export SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@253 -- # SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@254 -- # export SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@254 -- # SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@257 -- # export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@257 -- # QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@258 -- # export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@258 -- # VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@260 -- # export AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@260 -- # AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@263 -- # export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@263 -- # UNBIND_ENTIRE_IOMMU_GROUP=yes 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@265 -- # _LCOV_MAIN=0 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@266 -- # _LCOV_LLVM=1 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@267 -- # _LCOV= 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@268 -- # [[ '' == *clang* ]] 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@268 -- # [[ 0 -eq 1 ]] 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@270 -- # _lcov_opt[_LCOV_LLVM]='--gcov-tool /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/fuzz/llvm/llvm-gcov.sh' 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@271 -- # _lcov_opt[_LCOV_MAIN]= 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@273 -- # lcov_opt= 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@276 -- # '[' 0 -eq 0 ']' 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@277 -- # export valgrind= 00:14:20.129 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@277 -- # valgrind= 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@283 -- # uname -s 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@283 -- # '[' Linux = Linux ']' 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@284 -- # HUGEMEM=4096 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@285 -- # export CLEAR_HUGE=yes 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@285 -- # CLEAR_HUGE=yes 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@287 -- # MAKE=make 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@288 -- # MAKEFLAGS=-j144 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@304 -- # export HUGEMEM=4096 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@304 -- # HUGEMEM=4096 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@306 -- # NO_HUGE=() 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@307 -- # TEST_MODE= 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@308 -- # for i in "$@" 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@309 -- # case "$i" in 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@314 -- # TEST_TRANSPORT=tcp 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@329 -- # [[ -z 1768962 ]] 00:14:20.130 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@329 -- # kill -0 1768962 00:14:20.391 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1674 -- # set_test_storage 2147483648 00:14:20.391 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@339 -- # [[ -v testdir ]] 00:14:20.391 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@341 -- # local requested_size=2147483648 00:14:20.391 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@342 -- # local mount target_dir 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@344 -- # local -A mounts fss sizes avails uses 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@345 -- # local source fs size avail mount use 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@347 -- # local storage_fallback storage_candidates 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@349 -- # mktemp -udt spdk.XXXXXX 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@349 -- # storage_fallback=/tmp/spdk.uiwi5W 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@354 -- # storage_candidates=("$testdir" "$storage_fallback/tests/${testdir##*/}" "$storage_fallback") 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@356 -- # [[ -n '' ]] 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@361 -- # [[ -n '' ]] 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@366 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target /tmp/spdk.uiwi5W/tests/target /tmp/spdk.uiwi5W 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@369 -- # requested_size=2214592512 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@338 -- # df -T 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@338 -- # grep -v Filesystem 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=spdk_devtmpfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=devtmpfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=67108864 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=67108864 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=0 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=/dev/pmem0 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=ext2 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=4096 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=5284429824 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=5284425728 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=spdk_root 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=overlay 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=122527612928 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=129356541952 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=6828929024 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=tmpfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=tmpfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=64668237824 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=64678268928 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=10031104 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=tmpfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=tmpfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=25847947264 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=25871310848 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=23363584 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=efivarfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=efivarfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=216064 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=507904 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=287744 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=tmpfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=tmpfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=64677347328 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=64678273024 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=925696 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # mounts["$mount"]=tmpfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@372 -- # fss["$mount"]=tmpfs 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # avails["$mount"]=12935639040 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@373 -- # sizes["$mount"]=12935651328 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@374 -- # uses["$mount"]=12288 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@371 -- # read -r source fs size use avail _ mount 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@377 -- # printf '* Looking for test storage...\n' 00:14:20.392 * Looking for test storage... 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@379 -- # local target_space new_size 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@380 -- # for target_dir in "${storage_candidates[@]}" 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@383 -- # df /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@383 -- # awk '$1 !~ /Filesystem/{print $6}' 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@383 -- # mount=/ 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@385 -- # target_space=122527612928 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@386 -- # (( target_space == 0 || target_space < requested_size )) 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@389 -- # (( target_space >= requested_size )) 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@391 -- # [[ overlay == tmpfs ]] 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@391 -- # [[ overlay == ramfs ]] 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@391 -- # [[ / == / ]] 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@392 -- # new_size=9043521536 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@393 -- # (( new_size * 100 / sizes[/] > 95 )) 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@398 -- # export SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@398 -- # SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@399 -- # printf '* Found test storage at %s\n' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:20.392 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@400 -- # return 0 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1676 -- # set -o errtrace 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1677 -- # shopt -s extdebug 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1678 -- # trap 'trap - ERR; print_backtrace >&2' ERR 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1680 -- # PS4=' \t ${test_domain:-} -- ${BASH_SOURCE#${BASH_SOURCE%/*/*}/}@${LINENO} -- \$ ' 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1681 -- # true 00:14:20.392 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1683 -- # xtrace_fd 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@25 -- # [[ -n 15 ]] 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@25 -- # [[ -e /proc/self/fd/15 ]] 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@27 -- # exec 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@29 -- # exec 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@31 -- # xtrace_restore 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@16 -- # unset -v 'X_STACK[0 - 1 < 0 ? 0 : 0 - 1]' 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@17 -- # (( 0 == 0 )) 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@18 -- # set -x 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1689 -- # lcov --version 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # IFS=.-: 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@336 -- # read -ra ver1 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # IFS=.-: 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@337 -- # read -ra ver2 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@338 -- # local 'op=<' 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@340 -- # ver1_l=2 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@341 -- # ver2_l=1 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@344 -- # case "$op" in 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@345 -- # : 1 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # decimal 1 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=1 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 1 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@365 -- # ver1[v]=1 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # decimal 2 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@353 -- # local d=2 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@355 -- # echo 2 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@366 -- # ver2[v]=2 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@368 -- # return 0 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:14:20.393 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:20.393 --rc genhtml_branch_coverage=1 00:14:20.393 --rc genhtml_function_coverage=1 00:14:20.393 --rc genhtml_legend=1 00:14:20.393 --rc geninfo_all_blocks=1 00:14:20.393 --rc geninfo_unexecuted_blocks=1 00:14:20.393 00:14:20.393 ' 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:14:20.393 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:20.393 --rc genhtml_branch_coverage=1 00:14:20.393 --rc genhtml_function_coverage=1 00:14:20.393 --rc genhtml_legend=1 00:14:20.393 --rc geninfo_all_blocks=1 00:14:20.393 --rc geninfo_unexecuted_blocks=1 00:14:20.393 00:14:20.393 ' 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:14:20.393 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:20.393 --rc genhtml_branch_coverage=1 00:14:20.393 --rc genhtml_function_coverage=1 00:14:20.393 --rc genhtml_legend=1 00:14:20.393 --rc geninfo_all_blocks=1 00:14:20.393 --rc geninfo_unexecuted_blocks=1 00:14:20.393 00:14:20.393 ' 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:14:20.393 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:14:20.393 --rc genhtml_branch_coverage=1 00:14:20.393 --rc genhtml_function_coverage=1 00:14:20.393 --rc genhtml_legend=1 00:14:20.393 --rc geninfo_all_blocks=1 00:14:20.393 --rc geninfo_unexecuted_blocks=1 00:14:20.393 00:14:20.393 ' 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@7 -- # uname -s 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@15 -- # shopt -s extglob 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.393 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@5 -- # export PATH 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@51 -- # : 0 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:14:20.394 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@12 -- # MALLOC_BDEV_SIZE=512 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@15 -- # nvmftestinit 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@474 -- # prepare_net_devs 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@436 -- # local -g is_hw=no 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@438 -- # remove_spdk_ns 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@309 -- # xtrace_disable 00:14:20.394 08:50:10 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:14:28.541 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@315 -- # pci_devs=() 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@319 -- # net_devs=() 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@320 -- # e810=() 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@320 -- # local -ga e810 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@321 -- # x722=() 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@321 -- # local -ga x722 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@322 -- # mlx=() 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@322 -- # local -ga mlx 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:14:28.542 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:14:28.542 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:14:28.542 Found net devices under 0000:4b:00.0: cvl_0_0 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:14:28.542 Found net devices under 0000:4b:00.1: cvl_0_1 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@440 -- # is_hw=yes 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:14:28.542 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:14:28.543 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:28.543 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.585 ms 00:14:28.543 00:14:28.543 --- 10.0.0.2 ping statistics --- 00:14:28.543 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:28.543 rtt min/avg/max/mdev = 0.585/0.585/0.585/0.000 ms 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:28.543 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:28.543 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.258 ms 00:14:28.543 00:14:28.543 --- 10.0.0.1 ping statistics --- 00:14:28.543 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:28.543 rtt min/avg/max/mdev = 0.258/0.258/0.258/0.000 ms 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@448 -- # return 0 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@105 -- # run_test nvmf_filesystem_no_in_capsule nvmf_filesystem_part 0 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:14:28.543 ************************************ 00:14:28.543 START TEST nvmf_filesystem_no_in_capsule 00:14:28.543 ************************************ 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1125 -- # nvmf_filesystem_part 0 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@47 -- # in_capsule=0 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@724 -- # xtrace_disable 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@507 -- # nvmfpid=1772603 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@508 -- # waitforlisten 1772603 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@831 -- # '[' -z 1772603 ']' 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:28.543 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:28.543 08:50:17 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:28.543 [2024-11-06 08:50:17.807178] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:14:28.543 [2024-11-06 08:50:17.807241] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:28.543 [2024-11-06 08:50:17.885613] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:14:28.543 [2024-11-06 08:50:17.921102] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:28.543 [2024-11-06 08:50:17.921137] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:28.543 [2024-11-06 08:50:17.921145] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:28.543 [2024-11-06 08:50:17.921152] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:28.543 [2024-11-06 08:50:17.921158] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:28.543 [2024-11-06 08:50:17.924765] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:28.543 [2024-11-06 08:50:17.924940] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:14:28.543 [2024-11-06 08:50:17.925148] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:14:28.543 [2024-11-06 08:50:17.925148] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@864 -- # return 0 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@730 -- # xtrace_disable 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:28.543 [2024-11-06 08:50:18.053748] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:28.543 Malloc1 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:28.543 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:28.544 [2024-11-06 08:50:18.192774] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1378 -- # local bdev_name=Malloc1 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1379 -- # local bdev_info 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1380 -- # local bs 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1381 -- # local nb 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1382 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1382 -- # bdev_info='[ 00:14:28.544 { 00:14:28.544 "name": "Malloc1", 00:14:28.544 "aliases": [ 00:14:28.544 "ae344be4-571f-4a9f-873c-70818417f568" 00:14:28.544 ], 00:14:28.544 "product_name": "Malloc disk", 00:14:28.544 "block_size": 512, 00:14:28.544 "num_blocks": 1048576, 00:14:28.544 "uuid": "ae344be4-571f-4a9f-873c-70818417f568", 00:14:28.544 "assigned_rate_limits": { 00:14:28.544 "rw_ios_per_sec": 0, 00:14:28.544 "rw_mbytes_per_sec": 0, 00:14:28.544 "r_mbytes_per_sec": 0, 00:14:28.544 "w_mbytes_per_sec": 0 00:14:28.544 }, 00:14:28.544 "claimed": true, 00:14:28.544 "claim_type": "exclusive_write", 00:14:28.544 "zoned": false, 00:14:28.544 "supported_io_types": { 00:14:28.544 "read": true, 00:14:28.544 "write": true, 00:14:28.544 "unmap": true, 00:14:28.544 "flush": true, 00:14:28.544 "reset": true, 00:14:28.544 "nvme_admin": false, 00:14:28.544 "nvme_io": false, 00:14:28.544 "nvme_io_md": false, 00:14:28.544 "write_zeroes": true, 00:14:28.544 "zcopy": true, 00:14:28.544 "get_zone_info": false, 00:14:28.544 "zone_management": false, 00:14:28.544 "zone_append": false, 00:14:28.544 "compare": false, 00:14:28.544 "compare_and_write": false, 00:14:28.544 "abort": true, 00:14:28.544 "seek_hole": false, 00:14:28.544 "seek_data": false, 00:14:28.544 "copy": true, 00:14:28.544 "nvme_iov_md": false 00:14:28.544 }, 00:14:28.544 "memory_domains": [ 00:14:28.544 { 00:14:28.544 "dma_device_id": "system", 00:14:28.544 "dma_device_type": 1 00:14:28.544 }, 00:14:28.544 { 00:14:28.544 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:28.544 "dma_device_type": 2 00:14:28.544 } 00:14:28.544 ], 00:14:28.544 "driver_specific": {} 00:14:28.544 } 00:14:28.544 ]' 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1383 -- # jq '.[] .block_size' 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1383 -- # bs=512 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1384 -- # jq '.[] .num_blocks' 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1384 -- # nb=1048576 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1387 -- # bdev_size=512 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1388 -- # echo 512 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@58 -- # malloc_size=536870912 00:14:28.544 08:50:18 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:14:29.929 08:50:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:14:29.929 08:50:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1198 -- # local i=0 00:14:29.929 08:50:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:14:29.929 08:50:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:14:29.929 08:50:19 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1205 -- # sleep 2 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1208 -- # return 0 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@76 -- # local dev=nvme0n1 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@80 -- # echo 536870912 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@64 -- # nvme_size=536870912 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:14:31.844 08:50:21 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:14:32.105 08:50:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@69 -- # partprobe 00:14:33.048 08:50:22 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@70 -- # sleep 1 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@76 -- # '[' 0 -eq 0 ']' 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@77 -- # run_test filesystem_ext4 nvmf_filesystem_create ext4 nvme0n1 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:33.991 ************************************ 00:14:33.991 START TEST filesystem_ext4 00:14:33.991 ************************************ 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create ext4 nvme0n1 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@18 -- # fstype=ext4 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@926 -- # local fstype=ext4 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@928 -- # local i=0 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@929 -- # local force 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@931 -- # '[' ext4 = ext4 ']' 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@932 -- # force=-F 00:14:33.991 08:50:23 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@937 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:14:33.991 mke2fs 1.47.0 (5-Feb-2023) 00:14:33.991 Discarding device blocks: 0/522240 done 00:14:33.991 Creating filesystem with 522240 1k blocks and 130560 inodes 00:14:33.991 Filesystem UUID: 196aa5d7-5e1d-4868-b6f0-5136d6b75e11 00:14:33.991 Superblock backups stored on blocks: 00:14:33.991 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:14:33.991 00:14:33.991 Allocating group tables: 0/64 done 00:14:33.991 Writing inode tables: 0/64 done 00:14:36.538 Creating journal (8192 blocks): done 00:14:38.756 Writing superblocks and filesystem accounting information: 0/64 1/64 done 00:14:38.756 00:14:38.756 08:50:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@945 -- # return 0 00:14:38.756 08:50:28 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:14:45.523 08:50:34 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@25 -- # sync 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@27 -- # sync 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@29 -- # i=0 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@30 -- # umount /mnt/device 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@37 -- # kill -0 1772603 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:14:45.523 00:14:45.523 real 0m11.225s 00:14:45.523 user 0m0.030s 00:14:45.523 sys 0m0.082s 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@10 -- # set +x 00:14:45.523 ************************************ 00:14:45.523 END TEST filesystem_ext4 00:14:45.523 ************************************ 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@78 -- # run_test filesystem_btrfs nvmf_filesystem_create btrfs nvme0n1 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:45.523 ************************************ 00:14:45.523 START TEST filesystem_btrfs 00:14:45.523 ************************************ 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create btrfs nvme0n1 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@18 -- # fstype=btrfs 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@926 -- # local fstype=btrfs 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@928 -- # local i=0 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@929 -- # local force 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@931 -- # '[' btrfs = ext4 ']' 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@934 -- # force=-f 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@937 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:14:45.523 btrfs-progs v6.8.1 00:14:45.523 See https://btrfs.readthedocs.io for more information. 00:14:45.523 00:14:45.523 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:14:45.523 NOTE: several default settings have changed in version 5.15, please make sure 00:14:45.523 this does not affect your deployments: 00:14:45.523 - DUP for metadata (-m dup) 00:14:45.523 - enabled no-holes (-O no-holes) 00:14:45.523 - enabled free-space-tree (-R free-space-tree) 00:14:45.523 00:14:45.523 Label: (null) 00:14:45.523 UUID: 9476ea6c-8a0a-4897-93c0-69a3145b32a5 00:14:45.523 Node size: 16384 00:14:45.523 Sector size: 4096 (CPU page size: 4096) 00:14:45.523 Filesystem size: 510.00MiB 00:14:45.523 Block group profiles: 00:14:45.523 Data: single 8.00MiB 00:14:45.523 Metadata: DUP 32.00MiB 00:14:45.523 System: DUP 8.00MiB 00:14:45.523 SSD detected: yes 00:14:45.523 Zoned device: no 00:14:45.523 Features: extref, skinny-metadata, no-holes, free-space-tree 00:14:45.523 Checksum: crc32c 00:14:45.523 Number of devices: 1 00:14:45.523 Devices: 00:14:45.523 ID SIZE PATH 00:14:45.523 1 510.00MiB /dev/nvme0n1p1 00:14:45.523 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@945 -- # return 0 00:14:45.523 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:14:45.803 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:14:45.803 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@25 -- # sync 00:14:45.803 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:14:45.803 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@27 -- # sync 00:14:45.803 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@29 -- # i=0 00:14:45.803 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@37 -- # kill -0 1772603 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:14:46.063 00:14:46.063 real 0m0.783s 00:14:46.063 user 0m0.025s 00:14:46.063 sys 0m0.123s 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@10 -- # set +x 00:14:46.063 ************************************ 00:14:46.063 END TEST filesystem_btrfs 00:14:46.063 ************************************ 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@79 -- # run_test filesystem_xfs nvmf_filesystem_create xfs nvme0n1 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:46.063 08:50:35 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:46.063 ************************************ 00:14:46.063 START TEST filesystem_xfs 00:14:46.063 ************************************ 00:14:46.063 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create xfs nvme0n1 00:14:46.063 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@18 -- # fstype=xfs 00:14:46.063 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:14:46.063 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:14:46.063 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@926 -- # local fstype=xfs 00:14:46.063 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:14:46.063 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@928 -- # local i=0 00:14:46.064 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@929 -- # local force 00:14:46.064 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@931 -- # '[' xfs = ext4 ']' 00:14:46.064 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@934 -- # force=-f 00:14:46.064 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@937 -- # mkfs.xfs -f /dev/nvme0n1p1 00:14:46.064 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:14:46.064 = sectsz=512 attr=2, projid32bit=1 00:14:46.064 = crc=1 finobt=1, sparse=1, rmapbt=0 00:14:46.064 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:14:46.064 data = bsize=4096 blocks=130560, imaxpct=25 00:14:46.064 = sunit=0 swidth=0 blks 00:14:46.064 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:14:46.064 log =internal log bsize=4096 blocks=16384, version=2 00:14:46.064 = sectsz=512 sunit=0 blks, lazy-count=1 00:14:46.064 realtime =none extsz=4096 blocks=0, rtextents=0 00:14:47.006 Discarding blocks...Done. 00:14:47.006 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@945 -- # return 0 00:14:47.006 08:50:36 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:14:48.917 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:14:48.917 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@25 -- # sync 00:14:48.917 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:14:48.917 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@27 -- # sync 00:14:48.917 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@29 -- # i=0 00:14:48.917 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:14:48.917 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@37 -- # kill -0 1772603 00:14:48.918 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:14:48.918 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:14:48.918 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:14:48.918 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:14:48.918 00:14:48.918 real 0m2.924s 00:14:48.918 user 0m0.026s 00:14:48.918 sys 0m0.079s 00:14:48.918 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:48.918 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@10 -- # set +x 00:14:48.918 ************************************ 00:14:48.918 END TEST filesystem_xfs 00:14:48.918 ************************************ 00:14:48.918 08:50:38 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:14:49.178 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@93 -- # sync 00:14:49.178 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:14:49.179 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1219 -- # local i=0 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1231 -- # return 0 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@101 -- # killprocess 1772603 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@950 -- # '[' -z 1772603 ']' 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@954 -- # kill -0 1772603 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@955 -- # uname 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1772603 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1772603' 00:14:49.179 killing process with pid 1772603 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@969 -- # kill 1772603 00:14:49.179 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@974 -- # wait 1772603 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@102 -- # nvmfpid= 00:14:49.439 00:14:49.439 real 0m21.734s 00:14:49.439 user 1m25.914s 00:14:49.439 sys 0m1.436s 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1126 -- # xtrace_disable 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:49.439 ************************************ 00:14:49.439 END TEST nvmf_filesystem_no_in_capsule 00:14:49.439 ************************************ 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@106 -- # run_test nvmf_filesystem_in_capsule nvmf_filesystem_part 4096 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:14:49.439 ************************************ 00:14:49.439 START TEST nvmf_filesystem_in_capsule 00:14:49.439 ************************************ 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1125 -- # nvmf_filesystem_part 4096 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@47 -- # in_capsule=4096 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:14:49.439 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:14:49.703 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@724 -- # xtrace_disable 00:14:49.703 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:49.703 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@507 -- # nvmfpid=1777204 00:14:49.703 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@508 -- # waitforlisten 1777204 00:14:49.703 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:14:49.703 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@831 -- # '[' -z 1777204 ']' 00:14:49.703 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:49.703 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@836 -- # local max_retries=100 00:14:49.703 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:49.703 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:49.704 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@840 -- # xtrace_disable 00:14:49.704 08:50:39 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:49.704 [2024-11-06 08:50:39.627228] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:14:49.704 [2024-11-06 08:50:39.627309] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:49.704 [2024-11-06 08:50:39.710618] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:14:49.704 [2024-11-06 08:50:39.752768] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:49.704 [2024-11-06 08:50:39.752806] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:49.704 [2024-11-06 08:50:39.752814] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:49.704 [2024-11-06 08:50:39.752821] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:49.704 [2024-11-06 08:50:39.752827] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:49.704 [2024-11-06 08:50:39.754683] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:14:49.704 [2024-11-06 08:50:39.754786] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:14:49.704 [2024-11-06 08:50:39.754883] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:14:49.704 [2024-11-06 08:50:39.754883] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@864 -- # return 0 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@730 -- # xtrace_disable 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 4096 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:50.647 [2024-11-06 08:50:40.462781] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:50.647 Malloc1 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:50.647 [2024-11-06 08:50:40.603774] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1378 -- # local bdev_name=Malloc1 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1379 -- # local bdev_info 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1380 -- # local bs 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1381 -- # local nb 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1382 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:14:50.647 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1382 -- # bdev_info='[ 00:14:50.647 { 00:14:50.647 "name": "Malloc1", 00:14:50.647 "aliases": [ 00:14:50.647 "a5431d9e-2c20-4166-85bb-051e0fd21f93" 00:14:50.647 ], 00:14:50.647 "product_name": "Malloc disk", 00:14:50.648 "block_size": 512, 00:14:50.648 "num_blocks": 1048576, 00:14:50.648 "uuid": "a5431d9e-2c20-4166-85bb-051e0fd21f93", 00:14:50.648 "assigned_rate_limits": { 00:14:50.648 "rw_ios_per_sec": 0, 00:14:50.648 "rw_mbytes_per_sec": 0, 00:14:50.648 "r_mbytes_per_sec": 0, 00:14:50.648 "w_mbytes_per_sec": 0 00:14:50.648 }, 00:14:50.648 "claimed": true, 00:14:50.648 "claim_type": "exclusive_write", 00:14:50.648 "zoned": false, 00:14:50.648 "supported_io_types": { 00:14:50.648 "read": true, 00:14:50.648 "write": true, 00:14:50.648 "unmap": true, 00:14:50.648 "flush": true, 00:14:50.648 "reset": true, 00:14:50.648 "nvme_admin": false, 00:14:50.648 "nvme_io": false, 00:14:50.648 "nvme_io_md": false, 00:14:50.648 "write_zeroes": true, 00:14:50.648 "zcopy": true, 00:14:50.648 "get_zone_info": false, 00:14:50.648 "zone_management": false, 00:14:50.648 "zone_append": false, 00:14:50.648 "compare": false, 00:14:50.648 "compare_and_write": false, 00:14:50.648 "abort": true, 00:14:50.648 "seek_hole": false, 00:14:50.648 "seek_data": false, 00:14:50.648 "copy": true, 00:14:50.648 "nvme_iov_md": false 00:14:50.648 }, 00:14:50.648 "memory_domains": [ 00:14:50.648 { 00:14:50.648 "dma_device_id": "system", 00:14:50.648 "dma_device_type": 1 00:14:50.648 }, 00:14:50.648 { 00:14:50.648 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:14:50.648 "dma_device_type": 2 00:14:50.648 } 00:14:50.648 ], 00:14:50.648 "driver_specific": {} 00:14:50.648 } 00:14:50.648 ]' 00:14:50.648 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1383 -- # jq '.[] .block_size' 00:14:50.648 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1383 -- # bs=512 00:14:50.648 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1384 -- # jq '.[] .num_blocks' 00:14:50.648 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1384 -- # nb=1048576 00:14:50.648 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1387 -- # bdev_size=512 00:14:50.648 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1388 -- # echo 512 00:14:50.648 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@58 -- # malloc_size=536870912 00:14:50.648 08:50:40 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:14:52.559 08:50:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:14:52.559 08:50:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1198 -- # local i=0 00:14:52.559 08:50:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:14:52.559 08:50:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:14:52.559 08:50:42 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1205 -- # sleep 2 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1208 -- # return 0 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@76 -- # local dev=nvme0n1 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@80 -- # echo 536870912 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@64 -- # nvme_size=536870912 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:14:54.470 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:14:54.730 08:50:44 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@69 -- # partprobe 00:14:55.671 08:50:45 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@70 -- # sleep 1 00:14:56.612 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@76 -- # '[' 4096 -eq 0 ']' 00:14:56.612 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@81 -- # run_test filesystem_in_capsule_ext4 nvmf_filesystem_create ext4 nvme0n1 00:14:56.612 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:14:56.612 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:14:56.612 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:14:56.612 ************************************ 00:14:56.613 START TEST filesystem_in_capsule_ext4 00:14:56.613 ************************************ 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create ext4 nvme0n1 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@18 -- # fstype=ext4 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@926 -- # local fstype=ext4 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@928 -- # local i=0 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@929 -- # local force 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@931 -- # '[' ext4 = ext4 ']' 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@932 -- # force=-F 00:14:56.613 08:50:46 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@937 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:14:56.613 mke2fs 1.47.0 (5-Feb-2023) 00:14:56.613 Discarding device blocks: 0/522240 done 00:14:56.613 Creating filesystem with 522240 1k blocks and 130560 inodes 00:14:56.613 Filesystem UUID: 1a0b1bbd-76d9-479d-bce1-f6f353440e52 00:14:56.613 Superblock backups stored on blocks: 00:14:56.613 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:14:56.613 00:14:56.613 Allocating group tables: 0/64 done 00:14:56.613 Writing inode tables: 0/64 done 00:14:59.159 Creating journal (8192 blocks): done 00:14:59.159 Writing superblocks and filesystem accounting information: 0/64 done 00:14:59.159 00:14:59.159 08:50:49 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@945 -- # return 0 00:14:59.159 08:50:49 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@25 -- # sync 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@27 -- # sync 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@29 -- # i=0 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@30 -- # umount /mnt/device 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@37 -- # kill -0 1777204 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:15:05.806 08:50:54 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:15:05.806 00:15:05.806 real 0m8.465s 00:15:05.806 user 0m0.036s 00:15:05.806 sys 0m0.070s 00:15:05.806 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:05.806 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@10 -- # set +x 00:15:05.806 ************************************ 00:15:05.806 END TEST filesystem_in_capsule_ext4 00:15:05.806 ************************************ 00:15:05.806 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@82 -- # run_test filesystem_in_capsule_btrfs nvmf_filesystem_create btrfs nvme0n1 00:15:05.806 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:15:05.806 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:05.806 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:15:05.806 ************************************ 00:15:05.806 START TEST filesystem_in_capsule_btrfs 00:15:05.806 ************************************ 00:15:05.806 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create btrfs nvme0n1 00:15:05.806 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@18 -- # fstype=btrfs 00:15:05.806 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:15:05.807 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:15:05.807 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@926 -- # local fstype=btrfs 00:15:05.807 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:15:05.807 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@928 -- # local i=0 00:15:05.807 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@929 -- # local force 00:15:05.807 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@931 -- # '[' btrfs = ext4 ']' 00:15:05.807 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@934 -- # force=-f 00:15:05.807 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@937 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:15:05.807 btrfs-progs v6.8.1 00:15:05.807 See https://btrfs.readthedocs.io for more information. 00:15:05.807 00:15:05.807 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:15:05.807 NOTE: several default settings have changed in version 5.15, please make sure 00:15:05.807 this does not affect your deployments: 00:15:05.807 - DUP for metadata (-m dup) 00:15:05.807 - enabled no-holes (-O no-holes) 00:15:05.807 - enabled free-space-tree (-R free-space-tree) 00:15:05.807 00:15:05.807 Label: (null) 00:15:05.807 UUID: 0e1355f9-444f-45b6-af1b-be37f0768402 00:15:05.807 Node size: 16384 00:15:05.807 Sector size: 4096 (CPU page size: 4096) 00:15:05.807 Filesystem size: 510.00MiB 00:15:05.807 Block group profiles: 00:15:05.807 Data: single 8.00MiB 00:15:05.807 Metadata: DUP 32.00MiB 00:15:05.807 System: DUP 8.00MiB 00:15:05.807 SSD detected: yes 00:15:05.807 Zoned device: no 00:15:05.807 Features: extref, skinny-metadata, no-holes, free-space-tree 00:15:05.807 Checksum: crc32c 00:15:05.807 Number of devices: 1 00:15:05.807 Devices: 00:15:05.807 ID SIZE PATH 00:15:05.807 1 510.00MiB /dev/nvme0n1p1 00:15:05.807 00:15:05.807 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@945 -- # return 0 00:15:05.807 08:50:55 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@25 -- # sync 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@27 -- # sync 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@29 -- # i=0 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@37 -- # kill -0 1777204 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:15:06.382 00:15:06.382 real 0m1.292s 00:15:06.382 user 0m0.027s 00:15:06.382 sys 0m0.122s 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@10 -- # set +x 00:15:06.382 ************************************ 00:15:06.382 END TEST filesystem_in_capsule_btrfs 00:15:06.382 ************************************ 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@83 -- # run_test filesystem_in_capsule_xfs nvmf_filesystem_create xfs nvme0n1 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:15:06.382 ************************************ 00:15:06.382 START TEST filesystem_in_capsule_xfs 00:15:06.382 ************************************ 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@1125 -- # nvmf_filesystem_create xfs nvme0n1 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@18 -- # fstype=xfs 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@926 -- # local fstype=xfs 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@927 -- # local dev_name=/dev/nvme0n1p1 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@928 -- # local i=0 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@929 -- # local force 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@931 -- # '[' xfs = ext4 ']' 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@934 -- # force=-f 00:15:06.382 08:50:56 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@937 -- # mkfs.xfs -f /dev/nvme0n1p1 00:15:06.642 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:15:06.642 = sectsz=512 attr=2, projid32bit=1 00:15:06.642 = crc=1 finobt=1, sparse=1, rmapbt=0 00:15:06.642 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:15:06.642 data = bsize=4096 blocks=130560, imaxpct=25 00:15:06.642 = sunit=0 swidth=0 blks 00:15:06.642 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:15:06.643 log =internal log bsize=4096 blocks=16384, version=2 00:15:06.643 = sectsz=512 sunit=0 blks, lazy-count=1 00:15:06.643 realtime =none extsz=4096 blocks=0, rtextents=0 00:15:07.585 Discarding blocks...Done. 00:15:07.585 08:50:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@945 -- # return 0 00:15:07.585 08:50:57 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:15:10.128 08:50:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:15:10.128 08:50:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@25 -- # sync 00:15:10.128 08:50:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:15:10.128 08:50:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@27 -- # sync 00:15:10.128 08:50:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@29 -- # i=0 00:15:10.128 08:50:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:15:10.128 08:50:59 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@37 -- # kill -0 1777204 00:15:10.128 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:15:10.128 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:15:10.128 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:15:10.128 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:15:10.128 00:15:10.128 real 0m3.566s 00:15:10.128 user 0m0.027s 00:15:10.128 sys 0m0.078s 00:15:10.128 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:10.128 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@10 -- # set +x 00:15:10.128 ************************************ 00:15:10.128 END TEST filesystem_in_capsule_xfs 00:15:10.128 ************************************ 00:15:10.128 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:15:10.128 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@93 -- # sync 00:15:10.390 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:15:10.651 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1219 -- # local i=0 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1231 -- # return 0 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@101 -- # killprocess 1777204 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@950 -- # '[' -z 1777204 ']' 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@954 -- # kill -0 1777204 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@955 -- # uname 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1777204 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1777204' 00:15:10.651 killing process with pid 1777204 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@969 -- # kill 1777204 00:15:10.651 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@974 -- # wait 1777204 00:15:10.912 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@102 -- # nvmfpid= 00:15:10.912 00:15:10.912 real 0m21.366s 00:15:10.912 user 1m24.474s 00:15:10.912 sys 0m1.478s 00:15:10.912 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:10.912 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:15:10.912 ************************************ 00:15:10.912 END TEST nvmf_filesystem_in_capsule 00:15:10.912 ************************************ 00:15:10.912 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- target/filesystem.sh@108 -- # nvmftestfini 00:15:10.912 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@514 -- # nvmfcleanup 00:15:10.912 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@121 -- # sync 00:15:10.912 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:15:10.912 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@124 -- # set +e 00:15:10.912 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:15:10.912 08:51:00 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:15:10.912 rmmod nvme_tcp 00:15:10.912 rmmod nvme_fabrics 00:15:10.912 rmmod nvme_keyring 00:15:10.912 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@128 -- # set -e 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@129 -- # return 0 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@297 -- # iptr 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@789 -- # iptables-save 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@789 -- # iptables-restore 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:11.173 08:51:01 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:13.085 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:15:13.085 00:15:13.085 real 0m53.190s 00:15:13.085 user 2m52.674s 00:15:13.085 sys 0m8.669s 00:15:13.085 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:13.085 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:15:13.085 ************************************ 00:15:13.085 END TEST nvmf_filesystem 00:15:13.085 ************************************ 00:15:13.085 08:51:03 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@18 -- # run_test nvmf_target_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:15:13.085 08:51:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:15:13.085 08:51:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:13.085 08:51:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:15:13.348 ************************************ 00:15:13.348 START TEST nvmf_target_discovery 00:15:13.348 ************************************ 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:15:13.348 * Looking for test storage... 00:15:13.348 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1689 -- # lcov --version 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@336 -- # IFS=.-: 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@336 -- # read -ra ver1 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@337 -- # IFS=.-: 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@337 -- # read -ra ver2 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@338 -- # local 'op=<' 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@340 -- # ver1_l=2 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@341 -- # ver2_l=1 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@344 -- # case "$op" in 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@345 -- # : 1 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@365 -- # decimal 1 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@353 -- # local d=1 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@355 -- # echo 1 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@365 -- # ver1[v]=1 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@366 -- # decimal 2 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@353 -- # local d=2 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@355 -- # echo 2 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@366 -- # ver2[v]=2 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@368 -- # return 0 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:15:13.348 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:13.348 --rc genhtml_branch_coverage=1 00:15:13.348 --rc genhtml_function_coverage=1 00:15:13.348 --rc genhtml_legend=1 00:15:13.348 --rc geninfo_all_blocks=1 00:15:13.348 --rc geninfo_unexecuted_blocks=1 00:15:13.348 00:15:13.348 ' 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:15:13.348 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:13.348 --rc genhtml_branch_coverage=1 00:15:13.348 --rc genhtml_function_coverage=1 00:15:13.348 --rc genhtml_legend=1 00:15:13.348 --rc geninfo_all_blocks=1 00:15:13.348 --rc geninfo_unexecuted_blocks=1 00:15:13.348 00:15:13.348 ' 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:15:13.348 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:13.348 --rc genhtml_branch_coverage=1 00:15:13.348 --rc genhtml_function_coverage=1 00:15:13.348 --rc genhtml_legend=1 00:15:13.348 --rc geninfo_all_blocks=1 00:15:13.348 --rc geninfo_unexecuted_blocks=1 00:15:13.348 00:15:13.348 ' 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:15:13.348 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:13.348 --rc genhtml_branch_coverage=1 00:15:13.348 --rc genhtml_function_coverage=1 00:15:13.348 --rc genhtml_legend=1 00:15:13.348 --rc geninfo_all_blocks=1 00:15:13.348 --rc geninfo_unexecuted_blocks=1 00:15:13.348 00:15:13.348 ' 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@7 -- # uname -s 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@15 -- # shopt -s extglob 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:13.348 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@5 -- # export PATH 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@51 -- # : 0 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:13.349 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@11 -- # NULL_BDEV_SIZE=102400 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@12 -- # NULL_BLOCK_SIZE=512 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@13 -- # NVMF_PORT_REFERRAL=4430 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@15 -- # hash nvme 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@20 -- # nvmftestinit 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@474 -- # prepare_net_devs 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@436 -- # local -g is_hw=no 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@438 -- # remove_spdk_ns 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@309 -- # xtrace_disable 00:15:13.349 08:51:03 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@315 -- # pci_devs=() 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@315 -- # local -a pci_devs 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@316 -- # pci_net_devs=() 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@317 -- # pci_drivers=() 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@317 -- # local -A pci_drivers 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@319 -- # net_devs=() 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@319 -- # local -ga net_devs 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@320 -- # e810=() 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@320 -- # local -ga e810 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@321 -- # x722=() 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@321 -- # local -ga x722 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@322 -- # mlx=() 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@322 -- # local -ga mlx 00:15:21.489 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:15:21.490 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:15:21.490 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@416 -- # [[ up == up ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:15:21.490 Found net devices under 0000:4b:00.0: cvl_0_0 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@416 -- # [[ up == up ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:15:21.490 Found net devices under 0000:4b:00.1: cvl_0_1 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@440 -- # is_hw=yes 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:15:21.490 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:21.490 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.631 ms 00:15:21.490 00:15:21.490 --- 10.0.0.2 ping statistics --- 00:15:21.490 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:21.490 rtt min/avg/max/mdev = 0.631/0.631/0.631/0.000 ms 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:21.490 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:21.490 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.261 ms 00:15:21.490 00:15:21.490 --- 10.0.0.1 ping statistics --- 00:15:21.490 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:21.490 rtt min/avg/max/mdev = 0.261/0.261/0.261/0.000 ms 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@448 -- # return 0 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@21 -- # nvmfappstart -m 0xF 00:15:21.490 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@724 -- # xtrace_disable 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@507 -- # nvmfpid=1786276 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@508 -- # waitforlisten 1786276 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@831 -- # '[' -z 1786276 ']' 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:21.491 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:21.491 08:51:10 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.491 [2024-11-06 08:51:10.847490] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:15:21.491 [2024-11-06 08:51:10.847548] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:21.491 [2024-11-06 08:51:10.929247] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:21.491 [2024-11-06 08:51:10.968486] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:21.491 [2024-11-06 08:51:10.968525] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:21.491 [2024-11-06 08:51:10.968533] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:21.491 [2024-11-06 08:51:10.968539] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:21.491 [2024-11-06 08:51:10.968546] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:21.491 [2024-11-06 08:51:10.970361] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:15:21.491 [2024-11-06 08:51:10.970479] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:15:21.491 [2024-11-06 08:51:10.970634] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:21.491 [2024-11-06 08:51:10.970634] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@864 -- # return 0 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@730 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 [2024-11-06 08:51:11.693853] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # seq 1 4 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null1 102400 512 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 Null1 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Null1 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 [2024-11-06 08:51:11.754219] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null2 102400 512 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 Null2 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Null2 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null3 102400 512 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 Null3 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode3 -a -s SPDK00000000000003 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode3 Null3 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode3 -t tcp -a 10.0.0.2 -s 4420 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null4 102400 512 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:21.752 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.013 Null4 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode4 -a -s SPDK00000000000004 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode4 Null4 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode4 -t tcp -a 10.0.0.2 -s 4420 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@32 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.013 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@35 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 10.0.0.2 -s 4430 00:15:22.014 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.014 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.014 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.014 08:51:11 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@37 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 4420 00:15:22.274 00:15:22.274 Discovery Log Number of Records 6, Generation counter 6 00:15:22.274 =====Discovery Log Entry 0====== 00:15:22.274 trtype: tcp 00:15:22.274 adrfam: ipv4 00:15:22.274 subtype: current discovery subsystem 00:15:22.274 treq: not required 00:15:22.274 portid: 0 00:15:22.274 trsvcid: 4420 00:15:22.274 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:15:22.274 traddr: 10.0.0.2 00:15:22.274 eflags: explicit discovery connections, duplicate discovery information 00:15:22.274 sectype: none 00:15:22.274 =====Discovery Log Entry 1====== 00:15:22.274 trtype: tcp 00:15:22.274 adrfam: ipv4 00:15:22.274 subtype: nvme subsystem 00:15:22.274 treq: not required 00:15:22.274 portid: 0 00:15:22.274 trsvcid: 4420 00:15:22.274 subnqn: nqn.2016-06.io.spdk:cnode1 00:15:22.274 traddr: 10.0.0.2 00:15:22.274 eflags: none 00:15:22.274 sectype: none 00:15:22.274 =====Discovery Log Entry 2====== 00:15:22.274 trtype: tcp 00:15:22.274 adrfam: ipv4 00:15:22.274 subtype: nvme subsystem 00:15:22.274 treq: not required 00:15:22.274 portid: 0 00:15:22.274 trsvcid: 4420 00:15:22.274 subnqn: nqn.2016-06.io.spdk:cnode2 00:15:22.274 traddr: 10.0.0.2 00:15:22.274 eflags: none 00:15:22.274 sectype: none 00:15:22.274 =====Discovery Log Entry 3====== 00:15:22.274 trtype: tcp 00:15:22.274 adrfam: ipv4 00:15:22.274 subtype: nvme subsystem 00:15:22.274 treq: not required 00:15:22.274 portid: 0 00:15:22.274 trsvcid: 4420 00:15:22.274 subnqn: nqn.2016-06.io.spdk:cnode3 00:15:22.274 traddr: 10.0.0.2 00:15:22.274 eflags: none 00:15:22.274 sectype: none 00:15:22.274 =====Discovery Log Entry 4====== 00:15:22.274 trtype: tcp 00:15:22.274 adrfam: ipv4 00:15:22.274 subtype: nvme subsystem 00:15:22.274 treq: not required 00:15:22.274 portid: 0 00:15:22.274 trsvcid: 4420 00:15:22.274 subnqn: nqn.2016-06.io.spdk:cnode4 00:15:22.274 traddr: 10.0.0.2 00:15:22.274 eflags: none 00:15:22.274 sectype: none 00:15:22.274 =====Discovery Log Entry 5====== 00:15:22.274 trtype: tcp 00:15:22.274 adrfam: ipv4 00:15:22.274 subtype: discovery subsystem referral 00:15:22.274 treq: not required 00:15:22.274 portid: 0 00:15:22.274 trsvcid: 4430 00:15:22.274 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:15:22.274 traddr: 10.0.0.2 00:15:22.274 eflags: none 00:15:22.274 sectype: none 00:15:22.274 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@39 -- # echo 'Perform nvmf subsystem discovery via RPC' 00:15:22.274 Perform nvmf subsystem discovery via RPC 00:15:22.274 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@40 -- # rpc_cmd nvmf_get_subsystems 00:15:22.274 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.274 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.274 [ 00:15:22.274 { 00:15:22.274 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:15:22.274 "subtype": "Discovery", 00:15:22.274 "listen_addresses": [ 00:15:22.274 { 00:15:22.274 "trtype": "TCP", 00:15:22.274 "adrfam": "IPv4", 00:15:22.274 "traddr": "10.0.0.2", 00:15:22.274 "trsvcid": "4420" 00:15:22.274 } 00:15:22.274 ], 00:15:22.274 "allow_any_host": true, 00:15:22.274 "hosts": [] 00:15:22.274 }, 00:15:22.274 { 00:15:22.274 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:15:22.274 "subtype": "NVMe", 00:15:22.274 "listen_addresses": [ 00:15:22.274 { 00:15:22.274 "trtype": "TCP", 00:15:22.274 "adrfam": "IPv4", 00:15:22.274 "traddr": "10.0.0.2", 00:15:22.274 "trsvcid": "4420" 00:15:22.274 } 00:15:22.274 ], 00:15:22.274 "allow_any_host": true, 00:15:22.274 "hosts": [], 00:15:22.274 "serial_number": "SPDK00000000000001", 00:15:22.274 "model_number": "SPDK bdev Controller", 00:15:22.274 "max_namespaces": 32, 00:15:22.274 "min_cntlid": 1, 00:15:22.274 "max_cntlid": 65519, 00:15:22.274 "namespaces": [ 00:15:22.274 { 00:15:22.274 "nsid": 1, 00:15:22.274 "bdev_name": "Null1", 00:15:22.274 "name": "Null1", 00:15:22.274 "nguid": "2FBB29B20D064EB18E1EF4D0699505A9", 00:15:22.274 "uuid": "2fbb29b2-0d06-4eb1-8e1e-f4d0699505a9" 00:15:22.274 } 00:15:22.274 ] 00:15:22.274 }, 00:15:22.274 { 00:15:22.275 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:15:22.275 "subtype": "NVMe", 00:15:22.275 "listen_addresses": [ 00:15:22.275 { 00:15:22.275 "trtype": "TCP", 00:15:22.275 "adrfam": "IPv4", 00:15:22.275 "traddr": "10.0.0.2", 00:15:22.275 "trsvcid": "4420" 00:15:22.275 } 00:15:22.275 ], 00:15:22.275 "allow_any_host": true, 00:15:22.275 "hosts": [], 00:15:22.275 "serial_number": "SPDK00000000000002", 00:15:22.275 "model_number": "SPDK bdev Controller", 00:15:22.275 "max_namespaces": 32, 00:15:22.275 "min_cntlid": 1, 00:15:22.275 "max_cntlid": 65519, 00:15:22.275 "namespaces": [ 00:15:22.275 { 00:15:22.275 "nsid": 1, 00:15:22.275 "bdev_name": "Null2", 00:15:22.275 "name": "Null2", 00:15:22.275 "nguid": "7C4BF000ECDE436F979EE91D85EBF1D9", 00:15:22.275 "uuid": "7c4bf000-ecde-436f-979e-e91d85ebf1d9" 00:15:22.275 } 00:15:22.275 ] 00:15:22.275 }, 00:15:22.275 { 00:15:22.275 "nqn": "nqn.2016-06.io.spdk:cnode3", 00:15:22.275 "subtype": "NVMe", 00:15:22.275 "listen_addresses": [ 00:15:22.275 { 00:15:22.275 "trtype": "TCP", 00:15:22.275 "adrfam": "IPv4", 00:15:22.275 "traddr": "10.0.0.2", 00:15:22.275 "trsvcid": "4420" 00:15:22.275 } 00:15:22.275 ], 00:15:22.275 "allow_any_host": true, 00:15:22.275 "hosts": [], 00:15:22.275 "serial_number": "SPDK00000000000003", 00:15:22.275 "model_number": "SPDK bdev Controller", 00:15:22.275 "max_namespaces": 32, 00:15:22.275 "min_cntlid": 1, 00:15:22.275 "max_cntlid": 65519, 00:15:22.275 "namespaces": [ 00:15:22.275 { 00:15:22.275 "nsid": 1, 00:15:22.275 "bdev_name": "Null3", 00:15:22.275 "name": "Null3", 00:15:22.275 "nguid": "949E610C1FA44E2299B99A5172E41E61", 00:15:22.275 "uuid": "949e610c-1fa4-4e22-99b9-9a5172e41e61" 00:15:22.275 } 00:15:22.275 ] 00:15:22.275 }, 00:15:22.275 { 00:15:22.275 "nqn": "nqn.2016-06.io.spdk:cnode4", 00:15:22.275 "subtype": "NVMe", 00:15:22.275 "listen_addresses": [ 00:15:22.275 { 00:15:22.275 "trtype": "TCP", 00:15:22.275 "adrfam": "IPv4", 00:15:22.275 "traddr": "10.0.0.2", 00:15:22.275 "trsvcid": "4420" 00:15:22.275 } 00:15:22.275 ], 00:15:22.275 "allow_any_host": true, 00:15:22.275 "hosts": [], 00:15:22.275 "serial_number": "SPDK00000000000004", 00:15:22.275 "model_number": "SPDK bdev Controller", 00:15:22.275 "max_namespaces": 32, 00:15:22.275 "min_cntlid": 1, 00:15:22.275 "max_cntlid": 65519, 00:15:22.275 "namespaces": [ 00:15:22.275 { 00:15:22.275 "nsid": 1, 00:15:22.275 "bdev_name": "Null4", 00:15:22.275 "name": "Null4", 00:15:22.275 "nguid": "594ACE0210ED4EF580E428B8544BC725", 00:15:22.275 "uuid": "594ace02-10ed-4ef5-80e4-28b8544bc725" 00:15:22.275 } 00:15:22.275 ] 00:15:22.275 } 00:15:22.275 ] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # seq 1 4 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null1 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null2 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode3 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null3 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode4 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null4 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@47 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 10.0.0.2 -s 4430 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # rpc_cmd bdev_get_bdevs 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # jq -r '.[].name' 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@49 -- # check_bdevs= 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@50 -- # '[' -n '' ']' 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@55 -- # trap - SIGINT SIGTERM EXIT 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- target/discovery.sh@57 -- # nvmftestfini 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@514 -- # nvmfcleanup 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@121 -- # sync 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@124 -- # set +e 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@125 -- # for i in {1..20} 00:15:22.275 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:15:22.275 rmmod nvme_tcp 00:15:22.275 rmmod nvme_fabrics 00:15:22.275 rmmod nvme_keyring 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@128 -- # set -e 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@129 -- # return 0 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@515 -- # '[' -n 1786276 ']' 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@516 -- # killprocess 1786276 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@950 -- # '[' -z 1786276 ']' 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@954 -- # kill -0 1786276 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@955 -- # uname 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1786276 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1786276' 00:15:22.537 killing process with pid 1786276 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@969 -- # kill 1786276 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@974 -- # wait 1786276 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@297 -- # iptr 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@789 -- # iptables-save 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@789 -- # iptables-restore 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@302 -- # remove_spdk_ns 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:22.537 08:51:12 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:15:25.082 00:15:25.082 real 0m11.472s 00:15:25.082 user 0m8.854s 00:15:25.082 sys 0m5.905s 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:15:25.082 ************************************ 00:15:25.082 END TEST nvmf_target_discovery 00:15:25.082 ************************************ 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@19 -- # run_test nvmf_referrals /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:15:25.082 ************************************ 00:15:25.082 START TEST nvmf_referrals 00:15:25.082 ************************************ 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:15:25.082 * Looking for test storage... 00:15:25.082 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1689 -- # lcov --version 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@336 -- # IFS=.-: 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@336 -- # read -ra ver1 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@337 -- # IFS=.-: 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@337 -- # read -ra ver2 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@338 -- # local 'op=<' 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@340 -- # ver1_l=2 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@341 -- # ver2_l=1 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@344 -- # case "$op" in 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@345 -- # : 1 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@365 -- # decimal 1 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@353 -- # local d=1 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@355 -- # echo 1 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@365 -- # ver1[v]=1 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@366 -- # decimal 2 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@353 -- # local d=2 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@355 -- # echo 2 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@366 -- # ver2[v]=2 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@368 -- # return 0 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:15:25.082 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:25.082 --rc genhtml_branch_coverage=1 00:15:25.082 --rc genhtml_function_coverage=1 00:15:25.082 --rc genhtml_legend=1 00:15:25.082 --rc geninfo_all_blocks=1 00:15:25.082 --rc geninfo_unexecuted_blocks=1 00:15:25.082 00:15:25.082 ' 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:15:25.082 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:25.082 --rc genhtml_branch_coverage=1 00:15:25.082 --rc genhtml_function_coverage=1 00:15:25.082 --rc genhtml_legend=1 00:15:25.082 --rc geninfo_all_blocks=1 00:15:25.082 --rc geninfo_unexecuted_blocks=1 00:15:25.082 00:15:25.082 ' 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:15:25.082 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:25.082 --rc genhtml_branch_coverage=1 00:15:25.082 --rc genhtml_function_coverage=1 00:15:25.082 --rc genhtml_legend=1 00:15:25.082 --rc geninfo_all_blocks=1 00:15:25.082 --rc geninfo_unexecuted_blocks=1 00:15:25.082 00:15:25.082 ' 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:15:25.082 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:25.082 --rc genhtml_branch_coverage=1 00:15:25.082 --rc genhtml_function_coverage=1 00:15:25.082 --rc genhtml_legend=1 00:15:25.082 --rc geninfo_all_blocks=1 00:15:25.082 --rc geninfo_unexecuted_blocks=1 00:15:25.082 00:15:25.082 ' 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@7 -- # uname -s 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:25.082 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@15 -- # shopt -s extglob 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@5 -- # export PATH 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@51 -- # : 0 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:25.083 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@11 -- # NVMF_REFERRAL_IP_1=127.0.0.2 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@12 -- # NVMF_REFERRAL_IP_2=127.0.0.3 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@13 -- # NVMF_REFERRAL_IP_3=127.0.0.4 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@14 -- # NVMF_PORT_REFERRAL=4430 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@15 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@16 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@37 -- # nvmftestinit 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@474 -- # prepare_net_devs 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@436 -- # local -g is_hw=no 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@438 -- # remove_spdk_ns 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@309 -- # xtrace_disable 00:15:25.083 08:51:14 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@315 -- # pci_devs=() 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@315 -- # local -a pci_devs 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@316 -- # pci_net_devs=() 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@317 -- # pci_drivers=() 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@317 -- # local -A pci_drivers 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@319 -- # net_devs=() 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@319 -- # local -ga net_devs 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@320 -- # e810=() 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@320 -- # local -ga e810 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@321 -- # x722=() 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@321 -- # local -ga x722 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@322 -- # mlx=() 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@322 -- # local -ga mlx 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:33.328 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:15:33.329 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:15:33.329 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@416 -- # [[ up == up ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:15:33.329 Found net devices under 0000:4b:00.0: cvl_0_0 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@416 -- # [[ up == up ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:15:33.329 Found net devices under 0000:4b:00.1: cvl_0_1 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@440 -- # is_hw=yes 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:15:33.329 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:33.329 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.693 ms 00:15:33.329 00:15:33.329 --- 10.0.0.2 ping statistics --- 00:15:33.329 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:33.329 rtt min/avg/max/mdev = 0.693/0.693/0.693/0.000 ms 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:33.329 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:33.329 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.258 ms 00:15:33.329 00:15:33.329 --- 10.0.0.1 ping statistics --- 00:15:33.329 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:33.329 rtt min/avg/max/mdev = 0.258/0.258/0.258/0.000 ms 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@448 -- # return 0 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@38 -- # nvmfappstart -m 0xF 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@724 -- # xtrace_disable 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@507 -- # nvmfpid=1790922 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@508 -- # waitforlisten 1790922 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@831 -- # '[' -z 1790922 ']' 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:33.329 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:33.330 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:33.330 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:33.330 08:51:22 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.330 [2024-11-06 08:51:22.418708] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:15:33.330 [2024-11-06 08:51:22.418794] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:33.330 [2024-11-06 08:51:22.498559] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:33.330 [2024-11-06 08:51:22.535123] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:33.330 [2024-11-06 08:51:22.535153] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:33.330 [2024-11-06 08:51:22.535161] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:33.330 [2024-11-06 08:51:22.535168] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:33.330 [2024-11-06 08:51:22.535174] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:33.330 [2024-11-06 08:51:22.536731] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:15:33.330 [2024-11-06 08:51:22.536881] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:15:33.330 [2024-11-06 08:51:22.537094] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:33.330 [2024-11-06 08:51:22.537096] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@864 -- # return 0 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@730 -- # xtrace_disable 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@40 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.330 [2024-11-06 08:51:23.280234] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 10.0.0.2 -s 8009 discovery 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.330 [2024-11-06 08:51:23.303894] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@44 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@45 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.3 -s 4430 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@46 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.4 -s 4430 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # jq length 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@48 -- # (( 3 == 3 )) 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@49 -- # get_referral_ips rpc 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:15:33.330 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@49 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@50 -- # get_referral_ips nvme 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@50 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@52 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@53 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.3 -s 4430 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.591 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@54 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.4 -s 4430 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # jq length 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@56 -- # (( 0 == 0 )) 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@57 -- # get_referral_ips nvme 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:15:33.592 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:15:33.852 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 00:15:33.852 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@57 -- # [[ '' == '' ]] 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@60 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n discovery 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@62 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@65 -- # get_referral_ips rpc 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.2 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@65 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@66 -- # get_referral_ips nvme 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:15:33.853 08:51:23 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:15:34.113 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.2 00:15:34.113 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@66 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:15:34.113 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # get_discovery_entries 'nvme subsystem' 00:15:34.113 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # jq -r .subnqn 00:15:34.113 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:15:34.113 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:34.113 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@67 -- # [[ nqn.2016-06.io.spdk:cnode1 == \n\q\n\.\2\0\1\6\-\0\6\.\i\o\.\s\p\d\k\:\c\n\o\d\e\1 ]] 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # get_discovery_entries 'discovery subsystem referral' 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # jq -r .subnqn 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@68 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@71 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@73 -- # get_referral_ips rpc 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:15:34.373 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@73 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@74 -- # get_referral_ips nvme 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@74 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # get_discovery_entries 'nvme subsystem' 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # jq -r .subnqn 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:34.634 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:15:34.895 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@75 -- # [[ '' == '' ]] 00:15:34.895 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # get_discovery_entries 'discovery subsystem referral' 00:15:34.895 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # jq -r .subnqn 00:15:34.895 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:15:34.895 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:34.895 08:51:24 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@76 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@79 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2014-08.org.nvmexpress.discovery 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # rpc_cmd nvmf_discovery_get_referrals 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # jq length 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@82 -- # (( 0 == 0 )) 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@83 -- # get_referral_ips nvme 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 8009 -o json 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:15:35.156 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@26 -- # echo 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@83 -- # [[ '' == '' ]] 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@85 -- # trap - SIGINT SIGTERM EXIT 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- target/referrals.sh@86 -- # nvmftestfini 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@514 -- # nvmfcleanup 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@121 -- # sync 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@124 -- # set +e 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@125 -- # for i in {1..20} 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:15:35.417 rmmod nvme_tcp 00:15:35.417 rmmod nvme_fabrics 00:15:35.417 rmmod nvme_keyring 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@128 -- # set -e 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@129 -- # return 0 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@515 -- # '[' -n 1790922 ']' 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@516 -- # killprocess 1790922 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@950 -- # '[' -z 1790922 ']' 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@954 -- # kill -0 1790922 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@955 -- # uname 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1790922 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1790922' 00:15:35.417 killing process with pid 1790922 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@969 -- # kill 1790922 00:15:35.417 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@974 -- # wait 1790922 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@297 -- # iptr 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@789 -- # iptables-save 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@789 -- # iptables-restore 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@302 -- # remove_spdk_ns 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:35.679 08:51:25 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:37.592 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:15:37.592 00:15:37.592 real 0m12.902s 00:15:37.592 user 0m15.249s 00:15:37.592 sys 0m6.334s 00:15:37.592 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@1126 -- # xtrace_disable 00:15:37.592 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:15:37.592 ************************************ 00:15:37.592 END TEST nvmf_referrals 00:15:37.592 ************************************ 00:15:37.592 08:51:27 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@20 -- # run_test nvmf_connect_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:15:37.592 08:51:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:15:37.592 08:51:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:15:37.853 08:51:27 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:15:37.853 ************************************ 00:15:37.853 START TEST nvmf_connect_disconnect 00:15:37.853 ************************************ 00:15:37.853 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:15:37.853 * Looking for test storage... 00:15:37.853 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:37.853 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1689 -- # lcov --version 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@333 -- # local ver1 ver1_l 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@334 -- # local ver2 ver2_l 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@336 -- # IFS=.-: 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@336 -- # read -ra ver1 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@337 -- # IFS=.-: 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@337 -- # read -ra ver2 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@338 -- # local 'op=<' 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@340 -- # ver1_l=2 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@341 -- # ver2_l=1 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@344 -- # case "$op" in 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@345 -- # : 1 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@364 -- # (( v = 0 )) 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@365 -- # decimal 1 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@353 -- # local d=1 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@355 -- # echo 1 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@365 -- # ver1[v]=1 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@366 -- # decimal 2 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@353 -- # local d=2 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@355 -- # echo 2 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@366 -- # ver2[v]=2 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@368 -- # return 0 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:15:37.854 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:37.854 --rc genhtml_branch_coverage=1 00:15:37.854 --rc genhtml_function_coverage=1 00:15:37.854 --rc genhtml_legend=1 00:15:37.854 --rc geninfo_all_blocks=1 00:15:37.854 --rc geninfo_unexecuted_blocks=1 00:15:37.854 00:15:37.854 ' 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:15:37.854 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:37.854 --rc genhtml_branch_coverage=1 00:15:37.854 --rc genhtml_function_coverage=1 00:15:37.854 --rc genhtml_legend=1 00:15:37.854 --rc geninfo_all_blocks=1 00:15:37.854 --rc geninfo_unexecuted_blocks=1 00:15:37.854 00:15:37.854 ' 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:15:37.854 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:37.854 --rc genhtml_branch_coverage=1 00:15:37.854 --rc genhtml_function_coverage=1 00:15:37.854 --rc genhtml_legend=1 00:15:37.854 --rc geninfo_all_blocks=1 00:15:37.854 --rc geninfo_unexecuted_blocks=1 00:15:37.854 00:15:37.854 ' 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:15:37.854 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:15:37.854 --rc genhtml_branch_coverage=1 00:15:37.854 --rc genhtml_function_coverage=1 00:15:37.854 --rc genhtml_legend=1 00:15:37.854 --rc geninfo_all_blocks=1 00:15:37.854 --rc geninfo_unexecuted_blocks=1 00:15:37.854 00:15:37.854 ' 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@7 -- # uname -s 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:37.854 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@15 -- # shopt -s extglob 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:38.116 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@5 -- # export PATH 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@51 -- # : 0 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:15:38.117 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@55 -- # have_pci_nics=0 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@11 -- # MALLOC_BDEV_SIZE=64 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@15 -- # nvmftestinit 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@474 -- # prepare_net_devs 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@436 -- # local -g is_hw=no 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@438 -- # remove_spdk_ns 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@309 -- # xtrace_disable 00:15:38.117 08:51:27 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@315 -- # pci_devs=() 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@315 -- # local -a pci_devs 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@316 -- # pci_net_devs=() 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@317 -- # pci_drivers=() 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@317 -- # local -A pci_drivers 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@319 -- # net_devs=() 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@319 -- # local -ga net_devs 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@320 -- # e810=() 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@320 -- # local -ga e810 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@321 -- # x722=() 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@321 -- # local -ga x722 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@322 -- # mlx=() 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@322 -- # local -ga mlx 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:15:46.263 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:46.263 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:15:46.264 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@416 -- # [[ up == up ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:15:46.264 Found net devices under 0000:4b:00.0: cvl_0_0 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@416 -- # [[ up == up ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:15:46.264 Found net devices under 0000:4b:00.1: cvl_0_1 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@440 -- # is_hw=yes 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:15:46.264 08:51:34 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:15:46.264 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:46.264 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.645 ms 00:15:46.264 00:15:46.264 --- 10.0.0.2 ping statistics --- 00:15:46.264 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:46.264 rtt min/avg/max/mdev = 0.645/0.645/0.645/0.000 ms 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:46.264 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:46.264 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.321 ms 00:15:46.264 00:15:46.264 --- 10.0.0.1 ping statistics --- 00:15:46.264 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:46.264 rtt min/avg/max/mdev = 0.321/0.321/0.321/0.000 ms 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@448 -- # return 0 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@16 -- # nvmfappstart -m 0xF 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@724 -- # xtrace_disable 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@507 -- # nvmfpid=1795840 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@508 -- # waitforlisten 1795840 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@831 -- # '[' -z 1795840 ']' 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@836 -- # local max_retries=100 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:46.264 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@840 -- # xtrace_disable 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:46.264 08:51:35 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:15:46.264 [2024-11-06 08:51:35.253598] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:15:46.264 [2024-11-06 08:51:35.253667] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:46.264 [2024-11-06 08:51:35.336150] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:46.264 [2024-11-06 08:51:35.377869] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:46.264 [2024-11-06 08:51:35.377906] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:46.264 [2024-11-06 08:51:35.377914] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:46.264 [2024-11-06 08:51:35.377921] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:46.264 [2024-11-06 08:51:35.377927] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:46.264 [2024-11-06 08:51:35.379518] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:15:46.264 [2024-11-06 08:51:35.379631] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:15:46.264 [2024-11-06 08:51:35.379800] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:15:46.265 [2024-11-06 08:51:35.379812] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@864 -- # return 0 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@730 -- # xtrace_disable 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:46.265 [2024-11-06 08:51:36.107604] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@20 -- # bdev=Malloc0 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@561 -- # xtrace_disable 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:15:46.265 [2024-11-06 08:51:36.182127] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@26 -- # '[' 0 -eq 1 ']' 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@31 -- # num_iterations=5 00:15:46.265 08:51:36 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@34 -- # set +x 00:15:50.467 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:53.765 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:57.066 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:01.272 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:04.574 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@43 -- # trap - SIGINT SIGTERM EXIT 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- target/connect_disconnect.sh@45 -- # nvmftestfini 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@514 -- # nvmfcleanup 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@121 -- # sync 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@124 -- # set +e 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@125 -- # for i in {1..20} 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:16:04.574 rmmod nvme_tcp 00:16:04.574 rmmod nvme_fabrics 00:16:04.574 rmmod nvme_keyring 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@128 -- # set -e 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@129 -- # return 0 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@515 -- # '[' -n 1795840 ']' 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@516 -- # killprocess 1795840 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@950 -- # '[' -z 1795840 ']' 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@954 -- # kill -0 1795840 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@955 -- # uname 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1795840 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1795840' 00:16:04.574 killing process with pid 1795840 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@969 -- # kill 1795840 00:16:04.574 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@974 -- # wait 1795840 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@297 -- # iptr 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@789 -- # iptables-save 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@789 -- # iptables-restore 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@302 -- # remove_spdk_ns 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:16:04.835 08:51:54 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:07.379 08:51:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:16:07.379 00:16:07.379 real 0m29.146s 00:16:07.379 user 1m19.391s 00:16:07.379 sys 0m6.923s 00:16:07.379 08:51:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:07.379 08:51:56 nvmf_tcp.nvmf_target_extra.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:16:07.379 ************************************ 00:16:07.379 END TEST nvmf_connect_disconnect 00:16:07.379 ************************************ 00:16:07.379 08:51:56 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@21 -- # run_test nvmf_multitarget /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:16:07.379 08:51:56 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:16:07.379 08:51:56 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:07.379 08:51:56 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:16:07.379 ************************************ 00:16:07.379 START TEST nvmf_multitarget 00:16:07.379 ************************************ 00:16:07.379 08:51:56 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:16:07.379 * Looking for test storage... 00:16:07.379 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1689 -- # lcov --version 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@336 -- # IFS=.-: 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@336 -- # read -ra ver1 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@337 -- # IFS=.-: 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@337 -- # read -ra ver2 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@338 -- # local 'op=<' 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@340 -- # ver1_l=2 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@341 -- # ver2_l=1 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@344 -- # case "$op" in 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@345 -- # : 1 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@365 -- # decimal 1 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@353 -- # local d=1 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@355 -- # echo 1 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@365 -- # ver1[v]=1 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@366 -- # decimal 2 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@353 -- # local d=2 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@355 -- # echo 2 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@366 -- # ver2[v]=2 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:07.379 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@368 -- # return 0 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:16:07.380 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:07.380 --rc genhtml_branch_coverage=1 00:16:07.380 --rc genhtml_function_coverage=1 00:16:07.380 --rc genhtml_legend=1 00:16:07.380 --rc geninfo_all_blocks=1 00:16:07.380 --rc geninfo_unexecuted_blocks=1 00:16:07.380 00:16:07.380 ' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:16:07.380 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:07.380 --rc genhtml_branch_coverage=1 00:16:07.380 --rc genhtml_function_coverage=1 00:16:07.380 --rc genhtml_legend=1 00:16:07.380 --rc geninfo_all_blocks=1 00:16:07.380 --rc geninfo_unexecuted_blocks=1 00:16:07.380 00:16:07.380 ' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:16:07.380 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:07.380 --rc genhtml_branch_coverage=1 00:16:07.380 --rc genhtml_function_coverage=1 00:16:07.380 --rc genhtml_legend=1 00:16:07.380 --rc geninfo_all_blocks=1 00:16:07.380 --rc geninfo_unexecuted_blocks=1 00:16:07.380 00:16:07.380 ' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:16:07.380 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:07.380 --rc genhtml_branch_coverage=1 00:16:07.380 --rc genhtml_function_coverage=1 00:16:07.380 --rc genhtml_legend=1 00:16:07.380 --rc geninfo_all_blocks=1 00:16:07.380 --rc geninfo_unexecuted_blocks=1 00:16:07.380 00:16:07.380 ' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@7 -- # uname -s 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@15 -- # shopt -s extglob 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@5 -- # export PATH 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@51 -- # : 0 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:16:07.380 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@55 -- # have_pci_nics=0 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@13 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@15 -- # nvmftestinit 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@474 -- # prepare_net_devs 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@436 -- # local -g is_hw=no 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@438 -- # remove_spdk_ns 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@309 -- # xtrace_disable 00:16:07.380 08:51:57 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@315 -- # pci_devs=() 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@315 -- # local -a pci_devs 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@316 -- # pci_net_devs=() 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@317 -- # pci_drivers=() 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@317 -- # local -A pci_drivers 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@319 -- # net_devs=() 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@319 -- # local -ga net_devs 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@320 -- # e810=() 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@320 -- # local -ga e810 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@321 -- # x722=() 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@321 -- # local -ga x722 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@322 -- # mlx=() 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@322 -- # local -ga mlx 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:16:15.527 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:16:15.527 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@416 -- # [[ up == up ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:16:15.527 Found net devices under 0000:4b:00.0: cvl_0_0 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@416 -- # [[ up == up ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:16:15.527 Found net devices under 0000:4b:00.1: cvl_0_1 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@440 -- # is_hw=yes 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:16:15.527 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:16:15.527 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.511 ms 00:16:15.527 00:16:15.527 --- 10.0.0.2 ping statistics --- 00:16:15.527 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:15.527 rtt min/avg/max/mdev = 0.511/0.511/0.511/0.000 ms 00:16:15.527 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:16:15.528 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:16:15.528 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.318 ms 00:16:15.528 00:16:15.528 --- 10.0.0.1 ping statistics --- 00:16:15.528 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:15.528 rtt min/avg/max/mdev = 0.318/0.318/0.318/0.000 ms 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@448 -- # return 0 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@16 -- # nvmfappstart -m 0xF 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@507 -- # nvmfpid=1803931 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@508 -- # waitforlisten 1803931 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@831 -- # '[' -z 1803931 ']' 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:15.528 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:15.528 08:52:04 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:16:15.528 [2024-11-06 08:52:04.559745] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:16:15.528 [2024-11-06 08:52:04.559836] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:15.528 [2024-11-06 08:52:04.642661] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:16:15.528 [2024-11-06 08:52:04.684388] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:16:15.528 [2024-11-06 08:52:04.684423] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:16:15.528 [2024-11-06 08:52:04.684431] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:16:15.528 [2024-11-06 08:52:04.684438] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:16:15.528 [2024-11-06 08:52:04.684444] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:16:15.528 [2024-11-06 08:52:04.686292] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:15.528 [2024-11-06 08:52:04.686416] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:16:15.528 [2024-11-06 08:52:04.686576] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:15.528 [2024-11-06 08:52:04.686577] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@864 -- # return 0 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@18 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # jq length 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@21 -- # '[' 1 '!=' 1 ']' 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_1 -s 32 00:16:15.528 "nvmf_tgt_1" 00:16:15.528 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_2 -s 32 00:16:15.788 "nvmf_tgt_2" 00:16:15.788 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:16:15.788 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # jq length 00:16:15.788 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@28 -- # '[' 3 '!=' 3 ']' 00:16:15.789 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_1 00:16:16.050 true 00:16:16.050 08:52:05 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_2 00:16:16.050 true 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # jq length 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@35 -- # '[' 1 '!=' 1 ']' 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- target/multitarget.sh@41 -- # nvmftestfini 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@514 -- # nvmfcleanup 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@121 -- # sync 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@124 -- # set +e 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@125 -- # for i in {1..20} 00:16:16.050 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:16:16.050 rmmod nvme_tcp 00:16:16.311 rmmod nvme_fabrics 00:16:16.311 rmmod nvme_keyring 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@128 -- # set -e 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@129 -- # return 0 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@515 -- # '[' -n 1803931 ']' 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@516 -- # killprocess 1803931 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@950 -- # '[' -z 1803931 ']' 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@954 -- # kill -0 1803931 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@955 -- # uname 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1803931 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1803931' 00:16:16.311 killing process with pid 1803931 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@969 -- # kill 1803931 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@974 -- # wait 1803931 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@297 -- # iptr 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@789 -- # iptables-save 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@789 -- # iptables-restore 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@302 -- # remove_spdk_ns 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:16:16.311 08:52:06 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:18.860 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:16:18.860 00:16:18.860 real 0m11.505s 00:16:18.860 user 0m9.801s 00:16:18.860 sys 0m5.962s 00:16:18.860 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:18.860 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:16:18.860 ************************************ 00:16:18.860 END TEST nvmf_multitarget 00:16:18.860 ************************************ 00:16:18.860 08:52:08 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@22 -- # run_test nvmf_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:16:18.860 08:52:08 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:16:18.860 08:52:08 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:18.860 08:52:08 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:16:18.860 ************************************ 00:16:18.860 START TEST nvmf_rpc 00:16:18.860 ************************************ 00:16:18.860 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:16:18.860 * Looking for test storage... 00:16:18.860 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:16:18.860 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:16:18.860 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1689 -- # lcov --version 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@336 -- # IFS=.-: 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@336 -- # read -ra ver1 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@337 -- # IFS=.-: 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@337 -- # read -ra ver2 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@338 -- # local 'op=<' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@340 -- # ver1_l=2 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@341 -- # ver2_l=1 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@344 -- # case "$op" in 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@345 -- # : 1 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@365 -- # decimal 1 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@353 -- # local d=1 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@355 -- # echo 1 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@365 -- # ver1[v]=1 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@366 -- # decimal 2 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@353 -- # local d=2 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@355 -- # echo 2 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@366 -- # ver2[v]=2 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@368 -- # return 0 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:16:18.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:18.861 --rc genhtml_branch_coverage=1 00:16:18.861 --rc genhtml_function_coverage=1 00:16:18.861 --rc genhtml_legend=1 00:16:18.861 --rc geninfo_all_blocks=1 00:16:18.861 --rc geninfo_unexecuted_blocks=1 00:16:18.861 00:16:18.861 ' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:16:18.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:18.861 --rc genhtml_branch_coverage=1 00:16:18.861 --rc genhtml_function_coverage=1 00:16:18.861 --rc genhtml_legend=1 00:16:18.861 --rc geninfo_all_blocks=1 00:16:18.861 --rc geninfo_unexecuted_blocks=1 00:16:18.861 00:16:18.861 ' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:16:18.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:18.861 --rc genhtml_branch_coverage=1 00:16:18.861 --rc genhtml_function_coverage=1 00:16:18.861 --rc genhtml_legend=1 00:16:18.861 --rc geninfo_all_blocks=1 00:16:18.861 --rc geninfo_unexecuted_blocks=1 00:16:18.861 00:16:18.861 ' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:16:18.861 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:18.861 --rc genhtml_branch_coverage=1 00:16:18.861 --rc genhtml_function_coverage=1 00:16:18.861 --rc genhtml_legend=1 00:16:18.861 --rc geninfo_all_blocks=1 00:16:18.861 --rc geninfo_unexecuted_blocks=1 00:16:18.861 00:16:18.861 ' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@7 -- # uname -s 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@15 -- # shopt -s extglob 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@5 -- # export PATH 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@51 -- # : 0 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:16:18.861 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@55 -- # have_pci_nics=0 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@11 -- # loops=5 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@23 -- # nvmftestinit 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@474 -- # prepare_net_devs 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@436 -- # local -g is_hw=no 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@438 -- # remove_spdk_ns 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:18.861 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:16:18.862 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:16:18.862 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@309 -- # xtrace_disable 00:16:18.862 08:52:08 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@315 -- # pci_devs=() 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@315 -- # local -a pci_devs 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@316 -- # pci_net_devs=() 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@317 -- # pci_drivers=() 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@317 -- # local -A pci_drivers 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@319 -- # net_devs=() 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@319 -- # local -ga net_devs 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@320 -- # e810=() 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@320 -- # local -ga e810 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@321 -- # x722=() 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@321 -- # local -ga x722 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@322 -- # mlx=() 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@322 -- # local -ga mlx 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:16:25.452 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:16:25.452 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@416 -- # [[ up == up ]] 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:16:25.452 Found net devices under 0000:4b:00.0: cvl_0_0 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:16:25.452 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@416 -- # [[ up == up ]] 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:16:25.453 Found net devices under 0000:4b:00.1: cvl_0_1 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@440 -- # is_hw=yes 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:16:25.453 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:16:25.715 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:16:25.715 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.637 ms 00:16:25.715 00:16:25.715 --- 10.0.0.2 ping statistics --- 00:16:25.715 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:25.715 rtt min/avg/max/mdev = 0.637/0.637/0.637/0.000 ms 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:16:25.715 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:16:25.715 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.293 ms 00:16:25.715 00:16:25.715 --- 10.0.0.1 ping statistics --- 00:16:25.715 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:16:25.715 rtt min/avg/max/mdev = 0.293/0.293/0.293/0.000 ms 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@448 -- # return 0 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@24 -- # nvmfappstart -m 0xF 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@724 -- # xtrace_disable 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@507 -- # nvmfpid=1808338 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@508 -- # waitforlisten 1808338 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@831 -- # '[' -z 1808338 ']' 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@836 -- # local max_retries=100 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:16:25.715 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@840 -- # xtrace_disable 00:16:25.715 08:52:15 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:25.976 [2024-11-06 08:52:15.864255] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:16:25.976 [2024-11-06 08:52:15.864327] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:16:25.976 [2024-11-06 08:52:15.947661] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:16:25.976 [2024-11-06 08:52:15.989539] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:16:25.976 [2024-11-06 08:52:15.989576] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:16:25.976 [2024-11-06 08:52:15.989584] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:16:25.976 [2024-11-06 08:52:15.989591] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:16:25.976 [2024-11-06 08:52:15.989597] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:16:25.976 [2024-11-06 08:52:15.991459] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:16:25.976 [2024-11-06 08:52:15.991580] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:16:25.976 [2024-11-06 08:52:15.991740] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:16:25.976 [2024-11-06 08:52:15.991740] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:16:26.919 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@864 -- # return 0 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@730 -- # xtrace_disable 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@26 -- # rpc_cmd nvmf_get_stats 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@26 -- # stats='{ 00:16:26.920 "tick_rate": 2400000000, 00:16:26.920 "poll_groups": [ 00:16:26.920 { 00:16:26.920 "name": "nvmf_tgt_poll_group_000", 00:16:26.920 "admin_qpairs": 0, 00:16:26.920 "io_qpairs": 0, 00:16:26.920 "current_admin_qpairs": 0, 00:16:26.920 "current_io_qpairs": 0, 00:16:26.920 "pending_bdev_io": 0, 00:16:26.920 "completed_nvme_io": 0, 00:16:26.920 "transports": [] 00:16:26.920 }, 00:16:26.920 { 00:16:26.920 "name": "nvmf_tgt_poll_group_001", 00:16:26.920 "admin_qpairs": 0, 00:16:26.920 "io_qpairs": 0, 00:16:26.920 "current_admin_qpairs": 0, 00:16:26.920 "current_io_qpairs": 0, 00:16:26.920 "pending_bdev_io": 0, 00:16:26.920 "completed_nvme_io": 0, 00:16:26.920 "transports": [] 00:16:26.920 }, 00:16:26.920 { 00:16:26.920 "name": "nvmf_tgt_poll_group_002", 00:16:26.920 "admin_qpairs": 0, 00:16:26.920 "io_qpairs": 0, 00:16:26.920 "current_admin_qpairs": 0, 00:16:26.920 "current_io_qpairs": 0, 00:16:26.920 "pending_bdev_io": 0, 00:16:26.920 "completed_nvme_io": 0, 00:16:26.920 "transports": [] 00:16:26.920 }, 00:16:26.920 { 00:16:26.920 "name": "nvmf_tgt_poll_group_003", 00:16:26.920 "admin_qpairs": 0, 00:16:26.920 "io_qpairs": 0, 00:16:26.920 "current_admin_qpairs": 0, 00:16:26.920 "current_io_qpairs": 0, 00:16:26.920 "pending_bdev_io": 0, 00:16:26.920 "completed_nvme_io": 0, 00:16:26.920 "transports": [] 00:16:26.920 } 00:16:26.920 ] 00:16:26.920 }' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@28 -- # jcount '.poll_groups[].name' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@14 -- # local 'filter=.poll_groups[].name' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@15 -- # jq '.poll_groups[].name' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@15 -- # wc -l 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@28 -- # (( 4 == 4 )) 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@29 -- # jq '.poll_groups[0].transports[0]' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@29 -- # [[ null == null ]] 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@31 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:26.920 [2024-11-06 08:52:16.827984] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@33 -- # rpc_cmd nvmf_get_stats 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@33 -- # stats='{ 00:16:26.920 "tick_rate": 2400000000, 00:16:26.920 "poll_groups": [ 00:16:26.920 { 00:16:26.920 "name": "nvmf_tgt_poll_group_000", 00:16:26.920 "admin_qpairs": 0, 00:16:26.920 "io_qpairs": 0, 00:16:26.920 "current_admin_qpairs": 0, 00:16:26.920 "current_io_qpairs": 0, 00:16:26.920 "pending_bdev_io": 0, 00:16:26.920 "completed_nvme_io": 0, 00:16:26.920 "transports": [ 00:16:26.920 { 00:16:26.920 "trtype": "TCP" 00:16:26.920 } 00:16:26.920 ] 00:16:26.920 }, 00:16:26.920 { 00:16:26.920 "name": "nvmf_tgt_poll_group_001", 00:16:26.920 "admin_qpairs": 0, 00:16:26.920 "io_qpairs": 0, 00:16:26.920 "current_admin_qpairs": 0, 00:16:26.920 "current_io_qpairs": 0, 00:16:26.920 "pending_bdev_io": 0, 00:16:26.920 "completed_nvme_io": 0, 00:16:26.920 "transports": [ 00:16:26.920 { 00:16:26.920 "trtype": "TCP" 00:16:26.920 } 00:16:26.920 ] 00:16:26.920 }, 00:16:26.920 { 00:16:26.920 "name": "nvmf_tgt_poll_group_002", 00:16:26.920 "admin_qpairs": 0, 00:16:26.920 "io_qpairs": 0, 00:16:26.920 "current_admin_qpairs": 0, 00:16:26.920 "current_io_qpairs": 0, 00:16:26.920 "pending_bdev_io": 0, 00:16:26.920 "completed_nvme_io": 0, 00:16:26.920 "transports": [ 00:16:26.920 { 00:16:26.920 "trtype": "TCP" 00:16:26.920 } 00:16:26.920 ] 00:16:26.920 }, 00:16:26.920 { 00:16:26.920 "name": "nvmf_tgt_poll_group_003", 00:16:26.920 "admin_qpairs": 0, 00:16:26.920 "io_qpairs": 0, 00:16:26.920 "current_admin_qpairs": 0, 00:16:26.920 "current_io_qpairs": 0, 00:16:26.920 "pending_bdev_io": 0, 00:16:26.920 "completed_nvme_io": 0, 00:16:26.920 "transports": [ 00:16:26.920 { 00:16:26.920 "trtype": "TCP" 00:16:26.920 } 00:16:26.920 ] 00:16:26.920 } 00:16:26.920 ] 00:16:26.920 }' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@35 -- # jsum '.poll_groups[].admin_qpairs' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@35 -- # (( 0 == 0 )) 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@36 -- # jsum '.poll_groups[].io_qpairs' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@36 -- # (( 0 == 0 )) 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@38 -- # '[' rdma == tcp ']' 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@46 -- # MALLOC_BDEV_SIZE=64 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@47 -- # MALLOC_BLOCK_SIZE=512 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@49 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:26.920 Malloc1 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@52 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:26.920 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.921 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:16:26.921 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.921 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:26.921 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.921 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@54 -- # rpc_cmd nvmf_subsystem_allow_any_host -d nqn.2016-06.io.spdk:cnode1 00:16:26.921 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.921 08:52:16 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@55 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:26.921 [2024-11-06 08:52:17.016125] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@58 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -a 10.0.0.2 -s 4420 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@650 -- # local es=0 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -a 10.0.0.2 -s 4420 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@638 -- # local arg=nvme 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # type -t nvme 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # type -P nvme 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # arg=/usr/sbin/nvme 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # [[ -x /usr/sbin/nvme ]] 00:16:26.921 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@653 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -a 10.0.0.2 -s 4420 00:16:27.181 [2024-11-06 08:52:17.053097] ctrlr.c: 823:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be' 00:16:27.181 Failed to write to /dev/nvme-fabrics: Input/output error 00:16:27.181 could not add new controller: failed to write to nvme-fabrics device 00:16:27.181 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@653 -- # es=1 00:16:27.181 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:27.181 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:27.181 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:27.181 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@61 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:16:27.181 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:27.181 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:27.181 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:27.181 08:52:17 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@62 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:16:28.564 08:52:18 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@63 -- # waitforserial SPDKISFASTANDAWESOME 00:16:28.564 08:52:18 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:16:28.564 08:52:18 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:16:28.564 08:52:18 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:16:28.564 08:52:18 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@64 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:16:31.109 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@65 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@68 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.109 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@69 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@650 -- # local es=0 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@652 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@638 -- # local arg=nvme 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # type -t nvme 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # type -P nvme 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # arg=/usr/sbin/nvme 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@644 -- # [[ -x /usr/sbin/nvme ]] 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@653 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:16:31.110 [2024-11-06 08:52:20.819049] ctrlr.c: 823:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be' 00:16:31.110 Failed to write to /dev/nvme-fabrics: Input/output error 00:16:31.110 could not add new controller: failed to write to nvme-fabrics device 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@653 -- # es=1 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@72 -- # rpc_cmd nvmf_subsystem_allow_any_host -e nqn.2016-06.io.spdk:cnode1 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:31.110 08:52:20 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@73 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:16:32.493 08:52:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@74 -- # waitforserial SPDKISFASTANDAWESOME 00:16:32.493 08:52:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:16:32.493 08:52:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:16:32.493 08:52:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:16:32.493 08:52:22 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@75 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:16:34.404 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@76 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:16:34.404 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@78 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # seq 1 5 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:34.664 [2024-11-06 08:52:24.551755] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:34.664 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.665 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:16:34.665 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.665 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:34.665 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.665 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:16:34.665 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:34.665 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:34.665 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:34.665 08:52:24 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:16:36.047 08:52:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:16:36.047 08:52:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:16:36.047 08:52:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:16:36.047 08:52:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:16:36.048 08:52:26 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:16:38.596 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:38.596 [2024-11-06 08:52:28.305614] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:38.596 08:52:28 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:16:39.981 08:52:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:16:39.981 08:52:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:16:39.981 08:52:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:16:39.981 08:52:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:16:39.981 08:52:29 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:16:41.893 08:52:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:16:41.893 08:52:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:16:41.893 08:52:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:16:41.893 08:52:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:16:41.893 08:52:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:16:41.893 08:52:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:16:41.893 08:52:31 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:16:42.154 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:42.154 [2024-11-06 08:52:32.105183] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:42.154 08:52:32 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:16:43.546 08:52:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:16:43.546 08:52:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:16:43.546 08:52:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:16:43.546 08:52:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:16:43.546 08:52:33 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:16:45.571 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:16:45.571 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:16:45.571 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:16:45.571 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:16:45.571 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:16:45.571 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:16:45.571 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:16:45.832 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:16:45.832 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:45.833 [2024-11-06 08:52:35.818968] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:45.833 08:52:35 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:16:47.218 08:52:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:16:47.218 08:52:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:16:47.218 08:52:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:16:47.218 08:52:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:16:47.218 08:52:37 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:16:49.764 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:49.764 [2024-11-06 08:52:39.523370] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:49.764 08:52:39 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:16:51.151 08:52:41 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:16:51.151 08:52:41 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:16:51.151 08:52:41 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:16:51.151 08:52:41 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:16:51.151 08:52:41 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:16:53.064 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:16:53.064 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:53.325 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:16:53.325 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:16:53.325 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:16:53.325 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # seq 1 5 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 [2024-11-06 08:52:43.248468] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 [2024-11-06 08:52:43.308603] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 [2024-11-06 08:52:43.368774] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.326 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 [2024-11-06 08:52:43.441000] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 [2024-11-06 08:52:43.509245] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@110 -- # rpc_cmd nvmf_get_stats 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@561 -- # xtrace_disable 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:16:53.587 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@110 -- # stats='{ 00:16:53.587 "tick_rate": 2400000000, 00:16:53.587 "poll_groups": [ 00:16:53.587 { 00:16:53.587 "name": "nvmf_tgt_poll_group_000", 00:16:53.587 "admin_qpairs": 0, 00:16:53.587 "io_qpairs": 224, 00:16:53.587 "current_admin_qpairs": 0, 00:16:53.587 "current_io_qpairs": 0, 00:16:53.587 "pending_bdev_io": 0, 00:16:53.587 "completed_nvme_io": 275, 00:16:53.587 "transports": [ 00:16:53.587 { 00:16:53.587 "trtype": "TCP" 00:16:53.587 } 00:16:53.587 ] 00:16:53.587 }, 00:16:53.587 { 00:16:53.587 "name": "nvmf_tgt_poll_group_001", 00:16:53.587 "admin_qpairs": 1, 00:16:53.587 "io_qpairs": 223, 00:16:53.587 "current_admin_qpairs": 0, 00:16:53.587 "current_io_qpairs": 0, 00:16:53.587 "pending_bdev_io": 0, 00:16:53.587 "completed_nvme_io": 517, 00:16:53.587 "transports": [ 00:16:53.587 { 00:16:53.587 "trtype": "TCP" 00:16:53.587 } 00:16:53.587 ] 00:16:53.587 }, 00:16:53.587 { 00:16:53.587 "name": "nvmf_tgt_poll_group_002", 00:16:53.587 "admin_qpairs": 6, 00:16:53.587 "io_qpairs": 218, 00:16:53.587 "current_admin_qpairs": 0, 00:16:53.587 "current_io_qpairs": 0, 00:16:53.587 "pending_bdev_io": 0, 00:16:53.587 "completed_nvme_io": 220, 00:16:53.587 "transports": [ 00:16:53.587 { 00:16:53.587 "trtype": "TCP" 00:16:53.587 } 00:16:53.587 ] 00:16:53.587 }, 00:16:53.587 { 00:16:53.587 "name": "nvmf_tgt_poll_group_003", 00:16:53.587 "admin_qpairs": 0, 00:16:53.587 "io_qpairs": 224, 00:16:53.587 "current_admin_qpairs": 0, 00:16:53.587 "current_io_qpairs": 0, 00:16:53.587 "pending_bdev_io": 0, 00:16:53.587 "completed_nvme_io": 227, 00:16:53.587 "transports": [ 00:16:53.587 { 00:16:53.588 "trtype": "TCP" 00:16:53.588 } 00:16:53.588 ] 00:16:53.588 } 00:16:53.588 ] 00:16:53.588 }' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@112 -- # jsum '.poll_groups[].admin_qpairs' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@112 -- # (( 7 > 0 )) 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@113 -- # jsum '.poll_groups[].io_qpairs' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@113 -- # (( 889 > 0 )) 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@115 -- # '[' rdma == tcp ']' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@121 -- # trap - SIGINT SIGTERM EXIT 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- target/rpc.sh@123 -- # nvmftestfini 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@514 -- # nvmfcleanup 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@121 -- # sync 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@124 -- # set +e 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@125 -- # for i in {1..20} 00:16:53.588 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:16:53.588 rmmod nvme_tcp 00:16:53.848 rmmod nvme_fabrics 00:16:53.848 rmmod nvme_keyring 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@128 -- # set -e 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@129 -- # return 0 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@515 -- # '[' -n 1808338 ']' 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@516 -- # killprocess 1808338 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@950 -- # '[' -z 1808338 ']' 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@954 -- # kill -0 1808338 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@955 -- # uname 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1808338 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1808338' 00:16:53.848 killing process with pid 1808338 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@969 -- # kill 1808338 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@974 -- # wait 1808338 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@297 -- # iptr 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@789 -- # iptables-restore 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@789 -- # iptables-save 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@302 -- # remove_spdk_ns 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:16:53.848 08:52:43 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:56.393 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:16:56.393 00:16:56.393 real 0m37.462s 00:16:56.393 user 1m53.399s 00:16:56.393 sys 0m7.631s 00:16:56.393 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:16:56.393 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:16:56.393 ************************************ 00:16:56.393 END TEST nvmf_rpc 00:16:56.393 ************************************ 00:16:56.393 08:52:46 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@23 -- # run_test nvmf_invalid /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:16:56.393 08:52:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:16:56.393 08:52:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:16:56.393 08:52:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:16:56.393 ************************************ 00:16:56.393 START TEST nvmf_invalid 00:16:56.393 ************************************ 00:16:56.393 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:16:56.393 * Looking for test storage... 00:16:56.393 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:16:56.393 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:16:56.393 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1689 -- # lcov --version 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@333 -- # local ver1 ver1_l 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@334 -- # local ver2 ver2_l 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@336 -- # IFS=.-: 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@336 -- # read -ra ver1 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@337 -- # IFS=.-: 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@337 -- # read -ra ver2 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@338 -- # local 'op=<' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@340 -- # ver1_l=2 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@341 -- # ver2_l=1 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@344 -- # case "$op" in 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@345 -- # : 1 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@364 -- # (( v = 0 )) 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@365 -- # decimal 1 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@353 -- # local d=1 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@355 -- # echo 1 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@365 -- # ver1[v]=1 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@366 -- # decimal 2 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@353 -- # local d=2 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@355 -- # echo 2 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@366 -- # ver2[v]=2 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@368 -- # return 0 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:16:56.394 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:56.394 --rc genhtml_branch_coverage=1 00:16:56.394 --rc genhtml_function_coverage=1 00:16:56.394 --rc genhtml_legend=1 00:16:56.394 --rc geninfo_all_blocks=1 00:16:56.394 --rc geninfo_unexecuted_blocks=1 00:16:56.394 00:16:56.394 ' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:16:56.394 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:56.394 --rc genhtml_branch_coverage=1 00:16:56.394 --rc genhtml_function_coverage=1 00:16:56.394 --rc genhtml_legend=1 00:16:56.394 --rc geninfo_all_blocks=1 00:16:56.394 --rc geninfo_unexecuted_blocks=1 00:16:56.394 00:16:56.394 ' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:16:56.394 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:56.394 --rc genhtml_branch_coverage=1 00:16:56.394 --rc genhtml_function_coverage=1 00:16:56.394 --rc genhtml_legend=1 00:16:56.394 --rc geninfo_all_blocks=1 00:16:56.394 --rc geninfo_unexecuted_blocks=1 00:16:56.394 00:16:56.394 ' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:16:56.394 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:16:56.394 --rc genhtml_branch_coverage=1 00:16:56.394 --rc genhtml_function_coverage=1 00:16:56.394 --rc genhtml_legend=1 00:16:56.394 --rc geninfo_all_blocks=1 00:16:56.394 --rc geninfo_unexecuted_blocks=1 00:16:56.394 00:16:56.394 ' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@7 -- # uname -s 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@15 -- # shopt -s extglob 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@5 -- # export PATH 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@51 -- # : 0 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:16:56.394 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@55 -- # have_pci_nics=0 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@11 -- # multi_target_rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@12 -- # rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:16:56.394 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@14 -- # target=foobar 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@16 -- # RANDOM=0 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@34 -- # nvmftestinit 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@474 -- # prepare_net_devs 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@436 -- # local -g is_hw=no 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@438 -- # remove_spdk_ns 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@309 -- # xtrace_disable 00:16:56.395 08:52:46 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@315 -- # pci_devs=() 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@315 -- # local -a pci_devs 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@316 -- # pci_net_devs=() 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@317 -- # pci_drivers=() 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@317 -- # local -A pci_drivers 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@319 -- # net_devs=() 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@319 -- # local -ga net_devs 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@320 -- # e810=() 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@320 -- # local -ga e810 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@321 -- # x722=() 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@321 -- # local -ga x722 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@322 -- # mlx=() 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@322 -- # local -ga mlx 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:17:04.536 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:17:04.536 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@416 -- # [[ up == up ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:17:04.536 Found net devices under 0000:4b:00.0: cvl_0_0 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@416 -- # [[ up == up ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:17:04.536 Found net devices under 0000:4b:00.1: cvl_0_1 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@440 -- # is_hw=yes 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:17:04.536 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:17:04.536 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:17:04.536 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.573 ms 00:17:04.536 00:17:04.537 --- 10.0.0.2 ping statistics --- 00:17:04.537 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:04.537 rtt min/avg/max/mdev = 0.573/0.573/0.573/0.000 ms 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:17:04.537 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:17:04.537 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.242 ms 00:17:04.537 00:17:04.537 --- 10.0.0.1 ping statistics --- 00:17:04.537 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:04.537 rtt min/avg/max/mdev = 0.242/0.242/0.242/0.000 ms 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@448 -- # return 0 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@35 -- # nvmfappstart -m 0xF 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@507 -- # nvmfpid=1818197 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@508 -- # waitforlisten 1818197 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@831 -- # '[' -z 1818197 ']' 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:04.537 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:04.537 08:52:53 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:17:04.537 [2024-11-06 08:52:53.769940] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:17:04.537 [2024-11-06 08:52:53.769990] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:04.537 [2024-11-06 08:52:53.847531] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:17:04.537 [2024-11-06 08:52:53.883135] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:17:04.537 [2024-11-06 08:52:53.883171] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:17:04.537 [2024-11-06 08:52:53.883179] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:17:04.537 [2024-11-06 08:52:53.883186] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:17:04.537 [2024-11-06 08:52:53.883192] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:17:04.537 [2024-11-06 08:52:53.884928] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:04.537 [2024-11-06 08:52:53.885040] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:17:04.537 [2024-11-06 08:52:53.885194] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:04.537 [2024-11-06 08:52:53.885195] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:17:04.537 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:04.537 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@864 -- # return 0 00:17:04.537 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:17:04.537 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:04.537 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:17:04.537 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:17:04.537 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@37 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:17:04.537 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -t foobar nqn.2016-06.io.spdk:cnode30316 00:17:04.798 [2024-11-06 08:52:54.760595] nvmf_rpc.c: 396:rpc_nvmf_create_subsystem: *ERROR*: Unable to find target foobar 00:17:04.798 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@40 -- # out='request: 00:17:04.798 { 00:17:04.798 "nqn": "nqn.2016-06.io.spdk:cnode30316", 00:17:04.798 "tgt_name": "foobar", 00:17:04.798 "method": "nvmf_create_subsystem", 00:17:04.798 "req_id": 1 00:17:04.798 } 00:17:04.798 Got JSON-RPC error response 00:17:04.798 response: 00:17:04.798 { 00:17:04.798 "code": -32603, 00:17:04.798 "message": "Unable to find target foobar" 00:17:04.798 }' 00:17:04.798 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@41 -- # [[ request: 00:17:04.798 { 00:17:04.798 "nqn": "nqn.2016-06.io.spdk:cnode30316", 00:17:04.798 "tgt_name": "foobar", 00:17:04.798 "method": "nvmf_create_subsystem", 00:17:04.798 "req_id": 1 00:17:04.798 } 00:17:04.798 Got JSON-RPC error response 00:17:04.798 response: 00:17:04.798 { 00:17:04.798 "code": -32603, 00:17:04.798 "message": "Unable to find target foobar" 00:17:04.798 } == *\U\n\a\b\l\e\ \t\o\ \f\i\n\d\ \t\a\r\g\e\t* ]] 00:17:04.798 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # echo -e '\x1f' 00:17:04.798 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s $'SPDKISFASTANDAWESOME\037' nqn.2016-06.io.spdk:cnode14776 00:17:05.058 [2024-11-06 08:52:54.949263] nvmf_rpc.c: 413:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode14776: invalid serial number 'SPDKISFASTANDAWESOME' 00:17:05.058 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@45 -- # out='request: 00:17:05.058 { 00:17:05.058 "nqn": "nqn.2016-06.io.spdk:cnode14776", 00:17:05.058 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:17:05.058 "method": "nvmf_create_subsystem", 00:17:05.058 "req_id": 1 00:17:05.058 } 00:17:05.058 Got JSON-RPC error response 00:17:05.058 response: 00:17:05.058 { 00:17:05.058 "code": -32602, 00:17:05.058 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:17:05.058 }' 00:17:05.058 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@46 -- # [[ request: 00:17:05.058 { 00:17:05.058 "nqn": "nqn.2016-06.io.spdk:cnode14776", 00:17:05.058 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:17:05.058 "method": "nvmf_create_subsystem", 00:17:05.058 "req_id": 1 00:17:05.058 } 00:17:05.058 Got JSON-RPC error response 00:17:05.058 response: 00:17:05.058 { 00:17:05.058 "code": -32602, 00:17:05.058 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:17:05.058 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:17:05.058 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # echo -e '\x1f' 00:17:05.058 08:52:54 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d $'SPDK_Controller\037' nqn.2016-06.io.spdk:cnode5887 00:17:05.058 [2024-11-06 08:52:55.137839] nvmf_rpc.c: 422:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode5887: invalid model number 'SPDK_Controller' 00:17:05.058 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@50 -- # out='request: 00:17:05.058 { 00:17:05.058 "nqn": "nqn.2016-06.io.spdk:cnode5887", 00:17:05.058 "model_number": "SPDK_Controller\u001f", 00:17:05.058 "method": "nvmf_create_subsystem", 00:17:05.058 "req_id": 1 00:17:05.058 } 00:17:05.058 Got JSON-RPC error response 00:17:05.058 response: 00:17:05.058 { 00:17:05.058 "code": -32602, 00:17:05.058 "message": "Invalid MN SPDK_Controller\u001f" 00:17:05.058 }' 00:17:05.058 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@51 -- # [[ request: 00:17:05.058 { 00:17:05.058 "nqn": "nqn.2016-06.io.spdk:cnode5887", 00:17:05.058 "model_number": "SPDK_Controller\u001f", 00:17:05.058 "method": "nvmf_create_subsystem", 00:17:05.058 "req_id": 1 00:17:05.058 } 00:17:05.058 Got JSON-RPC error response 00:17:05.058 response: 00:17:05.058 { 00:17:05.058 "code": -32602, 00:17:05.058 "message": "Invalid MN SPDK_Controller\u001f" 00:17:05.058 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@54 -- # gen_random_s 21 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@19 -- # local length=21 ll 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@21 -- # local chars 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@22 -- # local string 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll = 0 )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 65 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x41' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=A 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 106 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6a' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=j 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 86 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x56' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=V 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 36 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x24' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='$' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 70 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x46' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=F 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 122 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x7a' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=z 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 60 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3c' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='<' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 113 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x71' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=q 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 53 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x35' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=5 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 78 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4e' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=N 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 100 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x64' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=d 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 54 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x36' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=6 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 77 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4d' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=M 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 51 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x33' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=3 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 39 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x27' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=\' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 122 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x7a' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=z 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 38 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x26' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+='&' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 112 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x70' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=p 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 102 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x66' 00:17:05.320 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=f 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 75 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4b' 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=K 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 116 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x74' 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@25 -- # string+=t 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@28 -- # [[ A == \- ]] 00:17:05.321 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@31 -- # echo 'AjV$FzJl"E>T~f@GxSnbU{y6=zeirSOr<8' 00:17:05.844 08:52:55 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d ' ~OMf'\''j}a|{A>Jl"E>T~f@GxSnbU{y6=zeirSOr<8' nqn.2016-06.io.spdk:cnode29525 00:17:06.105 [2024-11-06 08:52:55.996571] nvmf_rpc.c: 422:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode29525: invalid model number ' ~OMf'j}a|{A>Jl"E>T~f@GxSnbU{y6=zeirSOr<8' 00:17:06.105 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@58 -- # out='request: 00:17:06.105 { 00:17:06.105 "nqn": "nqn.2016-06.io.spdk:cnode29525", 00:17:06.105 "model_number": " ~OMf'\''j}a|{A>Jl\"E>T~f@GxSnbU{y6=zeirSOr<8", 00:17:06.105 "method": "nvmf_create_subsystem", 00:17:06.105 "req_id": 1 00:17:06.105 } 00:17:06.105 Got JSON-RPC error response 00:17:06.105 response: 00:17:06.105 { 00:17:06.105 "code": -32602, 00:17:06.105 "message": "Invalid MN ~OMf'\''j}a|{A>Jl\"E>T~f@GxSnbU{y6=zeirSOr<8" 00:17:06.105 }' 00:17:06.105 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@59 -- # [[ request: 00:17:06.105 { 00:17:06.105 "nqn": "nqn.2016-06.io.spdk:cnode29525", 00:17:06.105 "model_number": " ~OMf'j}a|{A>Jl\"E>T~f@GxSnbU{y6=zeirSOr<8", 00:17:06.105 "method": "nvmf_create_subsystem", 00:17:06.105 "req_id": 1 00:17:06.105 } 00:17:06.105 Got JSON-RPC error response 00:17:06.105 response: 00:17:06.105 { 00:17:06.105 "code": -32602, 00:17:06.105 "message": "Invalid MN ~OMf'j}a|{A>Jl\"E>T~f@GxSnbU{y6=zeirSOr<8" 00:17:06.105 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:17:06.105 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport --trtype tcp 00:17:06.105 [2024-11-06 08:52:56.181239] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:17:06.105 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode -s SPDK001 -a 00:17:06.366 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@64 -- # [[ tcp == \T\C\P ]] 00:17:06.366 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # echo '' 00:17:06.366 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # head -n 1 00:17:06.366 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@67 -- # IP= 00:17:06.366 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode -t tcp -a '' -s 4421 00:17:06.627 [2024-11-06 08:52:56.563830] nvmf_rpc.c: 783:nvmf_rpc_listen_paused: *ERROR*: Unable to remove listener, rc -2 00:17:06.627 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@69 -- # out='request: 00:17:06.627 { 00:17:06.627 "nqn": "nqn.2016-06.io.spdk:cnode", 00:17:06.627 "listen_address": { 00:17:06.627 "trtype": "tcp", 00:17:06.627 "traddr": "", 00:17:06.627 "trsvcid": "4421" 00:17:06.627 }, 00:17:06.627 "method": "nvmf_subsystem_remove_listener", 00:17:06.627 "req_id": 1 00:17:06.627 } 00:17:06.627 Got JSON-RPC error response 00:17:06.627 response: 00:17:06.627 { 00:17:06.627 "code": -32602, 00:17:06.627 "message": "Invalid parameters" 00:17:06.627 }' 00:17:06.627 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@70 -- # [[ request: 00:17:06.627 { 00:17:06.627 "nqn": "nqn.2016-06.io.spdk:cnode", 00:17:06.627 "listen_address": { 00:17:06.627 "trtype": "tcp", 00:17:06.627 "traddr": "", 00:17:06.627 "trsvcid": "4421" 00:17:06.627 }, 00:17:06.627 "method": "nvmf_subsystem_remove_listener", 00:17:06.627 "req_id": 1 00:17:06.627 } 00:17:06.627 Got JSON-RPC error response 00:17:06.627 response: 00:17:06.627 { 00:17:06.627 "code": -32602, 00:17:06.627 "message": "Invalid parameters" 00:17:06.627 } != *\U\n\a\b\l\e\ \t\o\ \s\t\o\p\ \l\i\s\t\e\n\e\r\.* ]] 00:17:06.627 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode17349 -i 0 00:17:06.887 [2024-11-06 08:52:56.748382] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode17349: invalid cntlid range [0-65519] 00:17:06.887 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@73 -- # out='request: 00:17:06.887 { 00:17:06.887 "nqn": "nqn.2016-06.io.spdk:cnode17349", 00:17:06.887 "min_cntlid": 0, 00:17:06.887 "method": "nvmf_create_subsystem", 00:17:06.887 "req_id": 1 00:17:06.887 } 00:17:06.887 Got JSON-RPC error response 00:17:06.887 response: 00:17:06.887 { 00:17:06.887 "code": -32602, 00:17:06.887 "message": "Invalid cntlid range [0-65519]" 00:17:06.887 }' 00:17:06.887 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@74 -- # [[ request: 00:17:06.887 { 00:17:06.887 "nqn": "nqn.2016-06.io.spdk:cnode17349", 00:17:06.887 "min_cntlid": 0, 00:17:06.887 "method": "nvmf_create_subsystem", 00:17:06.887 "req_id": 1 00:17:06.887 } 00:17:06.887 Got JSON-RPC error response 00:17:06.887 response: 00:17:06.887 { 00:17:06.887 "code": -32602, 00:17:06.887 "message": "Invalid cntlid range [0-65519]" 00:17:06.887 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:17:06.887 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@75 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1691 -i 65520 00:17:06.887 [2024-11-06 08:52:56.932946] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode1691: invalid cntlid range [65520-65519] 00:17:06.887 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@75 -- # out='request: 00:17:06.887 { 00:17:06.887 "nqn": "nqn.2016-06.io.spdk:cnode1691", 00:17:06.887 "min_cntlid": 65520, 00:17:06.887 "method": "nvmf_create_subsystem", 00:17:06.887 "req_id": 1 00:17:06.887 } 00:17:06.887 Got JSON-RPC error response 00:17:06.887 response: 00:17:06.887 { 00:17:06.887 "code": -32602, 00:17:06.887 "message": "Invalid cntlid range [65520-65519]" 00:17:06.887 }' 00:17:06.887 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@76 -- # [[ request: 00:17:06.887 { 00:17:06.887 "nqn": "nqn.2016-06.io.spdk:cnode1691", 00:17:06.887 "min_cntlid": 65520, 00:17:06.887 "method": "nvmf_create_subsystem", 00:17:06.887 "req_id": 1 00:17:06.887 } 00:17:06.887 Got JSON-RPC error response 00:17:06.887 response: 00:17:06.887 { 00:17:06.887 "code": -32602, 00:17:06.887 "message": "Invalid cntlid range [65520-65519]" 00:17:06.887 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:17:06.887 08:52:56 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode19594 -I 0 00:17:07.147 [2024-11-06 08:52:57.117550] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode19594: invalid cntlid range [1-0] 00:17:07.147 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@77 -- # out='request: 00:17:07.147 { 00:17:07.147 "nqn": "nqn.2016-06.io.spdk:cnode19594", 00:17:07.147 "max_cntlid": 0, 00:17:07.147 "method": "nvmf_create_subsystem", 00:17:07.147 "req_id": 1 00:17:07.147 } 00:17:07.147 Got JSON-RPC error response 00:17:07.147 response: 00:17:07.147 { 00:17:07.147 "code": -32602, 00:17:07.147 "message": "Invalid cntlid range [1-0]" 00:17:07.147 }' 00:17:07.147 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@78 -- # [[ request: 00:17:07.147 { 00:17:07.147 "nqn": "nqn.2016-06.io.spdk:cnode19594", 00:17:07.147 "max_cntlid": 0, 00:17:07.147 "method": "nvmf_create_subsystem", 00:17:07.147 "req_id": 1 00:17:07.147 } 00:17:07.147 Got JSON-RPC error response 00:17:07.147 response: 00:17:07.147 { 00:17:07.147 "code": -32602, 00:17:07.147 "message": "Invalid cntlid range [1-0]" 00:17:07.147 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:17:07.147 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode6400 -I 65520 00:17:07.409 [2024-11-06 08:52:57.306118] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode6400: invalid cntlid range [1-65520] 00:17:07.409 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@79 -- # out='request: 00:17:07.409 { 00:17:07.409 "nqn": "nqn.2016-06.io.spdk:cnode6400", 00:17:07.409 "max_cntlid": 65520, 00:17:07.409 "method": "nvmf_create_subsystem", 00:17:07.409 "req_id": 1 00:17:07.409 } 00:17:07.409 Got JSON-RPC error response 00:17:07.409 response: 00:17:07.409 { 00:17:07.409 "code": -32602, 00:17:07.409 "message": "Invalid cntlid range [1-65520]" 00:17:07.409 }' 00:17:07.409 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@80 -- # [[ request: 00:17:07.409 { 00:17:07.409 "nqn": "nqn.2016-06.io.spdk:cnode6400", 00:17:07.409 "max_cntlid": 65520, 00:17:07.409 "method": "nvmf_create_subsystem", 00:17:07.409 "req_id": 1 00:17:07.409 } 00:17:07.409 Got JSON-RPC error response 00:17:07.409 response: 00:17:07.409 { 00:17:07.409 "code": -32602, 00:17:07.409 "message": "Invalid cntlid range [1-65520]" 00:17:07.409 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:17:07.409 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode17038 -i 6 -I 5 00:17:07.409 [2024-11-06 08:52:57.494704] nvmf_rpc.c: 434:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode17038: invalid cntlid range [6-5] 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@83 -- # out='request: 00:17:07.669 { 00:17:07.669 "nqn": "nqn.2016-06.io.spdk:cnode17038", 00:17:07.669 "min_cntlid": 6, 00:17:07.669 "max_cntlid": 5, 00:17:07.669 "method": "nvmf_create_subsystem", 00:17:07.669 "req_id": 1 00:17:07.669 } 00:17:07.669 Got JSON-RPC error response 00:17:07.669 response: 00:17:07.669 { 00:17:07.669 "code": -32602, 00:17:07.669 "message": "Invalid cntlid range [6-5]" 00:17:07.669 }' 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@84 -- # [[ request: 00:17:07.669 { 00:17:07.669 "nqn": "nqn.2016-06.io.spdk:cnode17038", 00:17:07.669 "min_cntlid": 6, 00:17:07.669 "max_cntlid": 5, 00:17:07.669 "method": "nvmf_create_subsystem", 00:17:07.669 "req_id": 1 00:17:07.669 } 00:17:07.669 Got JSON-RPC error response 00:17:07.669 response: 00:17:07.669 { 00:17:07.669 "code": -32602, 00:17:07.669 "message": "Invalid cntlid range [6-5]" 00:17:07.669 } == *\I\n\v\a\l\i\d\ \c\n\t\l\i\d\ \r\a\n\g\e* ]] 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target --name foobar 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@87 -- # out='request: 00:17:07.669 { 00:17:07.669 "name": "foobar", 00:17:07.669 "method": "nvmf_delete_target", 00:17:07.669 "req_id": 1 00:17:07.669 } 00:17:07.669 Got JSON-RPC error response 00:17:07.669 response: 00:17:07.669 { 00:17:07.669 "code": -32602, 00:17:07.669 "message": "The specified target doesn'\''t exist, cannot delete it." 00:17:07.669 }' 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@88 -- # [[ request: 00:17:07.669 { 00:17:07.669 "name": "foobar", 00:17:07.669 "method": "nvmf_delete_target", 00:17:07.669 "req_id": 1 00:17:07.669 } 00:17:07.669 Got JSON-RPC error response 00:17:07.669 response: 00:17:07.669 { 00:17:07.669 "code": -32602, 00:17:07.669 "message": "The specified target doesn't exist, cannot delete it." 00:17:07.669 } == *\T\h\e\ \s\p\e\c\i\f\i\e\d\ \t\a\r\g\e\t\ \d\o\e\s\n\'\t\ \e\x\i\s\t\,\ \c\a\n\n\o\t\ \d\e\l\e\t\e\ \i\t\.* ]] 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@90 -- # trap - SIGINT SIGTERM EXIT 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- target/invalid.sh@91 -- # nvmftestfini 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@514 -- # nvmfcleanup 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@121 -- # sync 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@124 -- # set +e 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@125 -- # for i in {1..20} 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:17:07.669 rmmod nvme_tcp 00:17:07.669 rmmod nvme_fabrics 00:17:07.669 rmmod nvme_keyring 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@128 -- # set -e 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@129 -- # return 0 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@515 -- # '[' -n 1818197 ']' 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@516 -- # killprocess 1818197 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@950 -- # '[' -z 1818197 ']' 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@954 -- # kill -0 1818197 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@955 -- # uname 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1818197 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1818197' 00:17:07.669 killing process with pid 1818197 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@969 -- # kill 1818197 00:17:07.669 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@974 -- # wait 1818197 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@297 -- # iptr 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@789 -- # iptables-save 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@789 -- # iptables-restore 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@302 -- # remove_spdk_ns 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:17:07.930 08:52:57 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:10.477 08:52:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:17:10.477 00:17:10.477 real 0m13.855s 00:17:10.477 user 0m20.508s 00:17:10.477 sys 0m6.483s 00:17:10.477 08:52:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:10.477 08:52:59 nvmf_tcp.nvmf_target_extra.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:17:10.477 ************************************ 00:17:10.477 END TEST nvmf_invalid 00:17:10.477 ************************************ 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@24 -- # run_test nvmf_connect_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:17:10.477 ************************************ 00:17:10.477 START TEST nvmf_connect_stress 00:17:10.477 ************************************ 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:17:10.477 * Looking for test storage... 00:17:10.477 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1689 -- # lcov --version 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@336 -- # IFS=.-: 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@336 -- # read -ra ver1 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@337 -- # IFS=.-: 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@337 -- # read -ra ver2 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@338 -- # local 'op=<' 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@340 -- # ver1_l=2 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@341 -- # ver2_l=1 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@344 -- # case "$op" in 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@345 -- # : 1 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@365 -- # decimal 1 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@353 -- # local d=1 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@355 -- # echo 1 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@366 -- # decimal 2 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@353 -- # local d=2 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@355 -- # echo 2 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@368 -- # return 0 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:10.477 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:17:10.478 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:10.478 --rc genhtml_branch_coverage=1 00:17:10.478 --rc genhtml_function_coverage=1 00:17:10.478 --rc genhtml_legend=1 00:17:10.478 --rc geninfo_all_blocks=1 00:17:10.478 --rc geninfo_unexecuted_blocks=1 00:17:10.478 00:17:10.478 ' 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:17:10.478 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:10.478 --rc genhtml_branch_coverage=1 00:17:10.478 --rc genhtml_function_coverage=1 00:17:10.478 --rc genhtml_legend=1 00:17:10.478 --rc geninfo_all_blocks=1 00:17:10.478 --rc geninfo_unexecuted_blocks=1 00:17:10.478 00:17:10.478 ' 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:17:10.478 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:10.478 --rc genhtml_branch_coverage=1 00:17:10.478 --rc genhtml_function_coverage=1 00:17:10.478 --rc genhtml_legend=1 00:17:10.478 --rc geninfo_all_blocks=1 00:17:10.478 --rc geninfo_unexecuted_blocks=1 00:17:10.478 00:17:10.478 ' 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:17:10.478 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:10.478 --rc genhtml_branch_coverage=1 00:17:10.478 --rc genhtml_function_coverage=1 00:17:10.478 --rc genhtml_legend=1 00:17:10.478 --rc geninfo_all_blocks=1 00:17:10.478 --rc geninfo_unexecuted_blocks=1 00:17:10.478 00:17:10.478 ' 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@7 -- # uname -s 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@5 -- # export PATH 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@51 -- # : 0 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:17:10.478 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@12 -- # nvmftestinit 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@474 -- # prepare_net_devs 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@436 -- # local -g is_hw=no 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@438 -- # remove_spdk_ns 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:17:10.478 08:53:00 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@319 -- # net_devs=() 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@320 -- # e810=() 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@320 -- # local -ga e810 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@321 -- # x722=() 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@321 -- # local -ga x722 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@322 -- # mlx=() 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:17:18.629 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:17:18.630 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:17:18.630 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:17:18.630 Found net devices under 0000:4b:00.0: cvl_0_0 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:17:18.630 Found net devices under 0000:4b:00.1: cvl_0_1 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@440 -- # is_hw=yes 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:17:18.630 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:17:18.630 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.619 ms 00:17:18.630 00:17:18.630 --- 10.0.0.2 ping statistics --- 00:17:18.630 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:18.630 rtt min/avg/max/mdev = 0.619/0.619/0.619/0.000 ms 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:17:18.630 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:17:18.630 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.255 ms 00:17:18.630 00:17:18.630 --- 10.0.0.1 ping statistics --- 00:17:18.630 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:18.630 rtt min/avg/max/mdev = 0.255/0.255/0.255/0.000 ms 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@448 -- # return 0 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@13 -- # nvmfappstart -m 0xE 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@507 -- # nvmfpid=1823381 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@508 -- # waitforlisten 1823381 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@831 -- # '[' -z 1823381 ']' 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:18.630 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:18.631 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:18.631 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:18.631 08:53:07 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:18.631 [2024-11-06 08:53:07.692657] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:17:18.631 [2024-11-06 08:53:07.692711] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:18.631 [2024-11-06 08:53:07.788832] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:17:18.631 [2024-11-06 08:53:07.829778] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:17:18.631 [2024-11-06 08:53:07.829822] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:17:18.631 [2024-11-06 08:53:07.829830] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:17:18.631 [2024-11-06 08:53:07.829838] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:17:18.631 [2024-11-06 08:53:07.829844] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:17:18.631 [2024-11-06 08:53:07.831396] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:17:18.631 [2024-11-06 08:53:07.831556] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:18.631 [2024-11-06 08:53:07.831557] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@864 -- # return 0 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:18.631 [2024-11-06 08:53:08.541733] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:18.631 [2024-11-06 08:53:08.566192] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:18.631 NULL1 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@21 -- # PERF_PID=1823419 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@23 -- # rpcs=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/connect_stress/connect_stress -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -t 10 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@25 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # seq 1 20 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:18.631 08:53:08 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:19.204 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:19.204 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:19.204 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:19.204 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:19.204 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:19.466 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:19.466 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:19.466 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:19.466 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:19.466 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:19.727 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:19.728 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:19.728 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:19.728 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:19.728 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:19.988 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:19.988 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:19.988 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:19.989 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:19.989 08:53:09 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:20.250 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:20.250 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:20.250 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:20.250 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:20.250 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:20.823 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:20.823 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:20.823 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:20.823 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:20.823 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:21.084 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:21.084 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:21.084 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:21.084 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:21.084 08:53:10 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:21.345 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:21.345 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:21.345 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:21.345 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:21.345 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:21.606 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:21.606 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:21.606 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:21.606 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:21.606 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:21.867 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:21.867 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:21.867 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:21.867 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:21.867 08:53:11 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:22.439 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:22.439 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:22.439 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:22.439 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:22.439 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:22.699 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:22.699 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:22.699 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:22.699 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:22.699 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:22.960 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:22.960 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:22.960 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:22.960 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:22.960 08:53:12 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:23.220 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:23.220 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:23.220 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:23.220 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:23.220 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:23.480 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:23.480 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:23.480 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:23.480 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:23.480 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:24.050 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:24.050 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:24.050 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:24.050 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:24.050 08:53:13 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:24.310 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:24.310 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:24.310 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:24.310 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:24.310 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:24.571 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:24.571 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:24.571 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:24.571 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:24.571 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:24.831 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:24.832 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:24.832 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:24.832 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:24.832 08:53:14 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:25.091 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:25.092 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:25.092 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:25.092 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:25.092 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:25.662 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:25.662 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:25.662 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:25.662 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:25.662 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:25.923 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:25.923 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:25.924 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:25.924 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:25.924 08:53:15 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:26.184 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.184 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:26.184 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:26.184 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.184 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:26.446 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.446 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:26.446 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:26.446 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.446 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:26.706 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:26.706 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:26.706 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:26.706 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:26.706 08:53:16 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:27.278 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:27.278 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:27.278 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:27.278 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:27.278 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:27.539 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:27.539 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:27.539 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:27.539 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:27.539 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:27.800 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:27.800 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:27.800 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:27.800 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:27.800 08:53:17 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:28.061 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:28.061 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:28.061 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:28.061 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:28.061 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:28.632 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:28.632 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:28.632 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:17:28.632 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:28.632 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:28.632 Testing NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 1823419 00:17:28.892 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh: line 34: kill: (1823419) - No such process 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@38 -- # wait 1823419 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@39 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- target/connect_stress.sh@43 -- # nvmftestfini 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@514 -- # nvmfcleanup 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@121 -- # sync 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@124 -- # set +e 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:17:28.892 rmmod nvme_tcp 00:17:28.892 rmmod nvme_fabrics 00:17:28.892 rmmod nvme_keyring 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@128 -- # set -e 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@129 -- # return 0 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@515 -- # '[' -n 1823381 ']' 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@516 -- # killprocess 1823381 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@950 -- # '[' -z 1823381 ']' 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@954 -- # kill -0 1823381 00:17:28.892 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@955 -- # uname 00:17:28.893 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:28.893 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1823381 00:17:28.893 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:17:28.893 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:17:28.893 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1823381' 00:17:28.893 killing process with pid 1823381 00:17:28.893 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@969 -- # kill 1823381 00:17:28.893 08:53:18 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@974 -- # wait 1823381 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@297 -- # iptr 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@789 -- # iptables-save 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@789 -- # iptables-restore 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:17:29.153 08:53:19 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:31.077 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:17:31.077 00:17:31.077 real 0m21.044s 00:17:31.077 user 0m42.246s 00:17:31.077 sys 0m8.930s 00:17:31.077 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:31.077 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:17:31.077 ************************************ 00:17:31.077 END TEST nvmf_connect_stress 00:17:31.077 ************************************ 00:17:31.077 08:53:21 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@25 -- # run_test nvmf_fused_ordering /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:17:31.077 08:53:21 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:17:31.077 08:53:21 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:31.077 08:53:21 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:17:31.077 ************************************ 00:17:31.077 START TEST nvmf_fused_ordering 00:17:31.077 ************************************ 00:17:31.077 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:17:31.338 * Looking for test storage... 00:17:31.338 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1689 -- # lcov --version 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@336 -- # IFS=.-: 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@336 -- # read -ra ver1 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@337 -- # IFS=.-: 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@337 -- # read -ra ver2 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@338 -- # local 'op=<' 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@340 -- # ver1_l=2 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@341 -- # ver2_l=1 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@344 -- # case "$op" in 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@345 -- # : 1 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@365 -- # decimal 1 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@353 -- # local d=1 00:17:31.338 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@355 -- # echo 1 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@365 -- # ver1[v]=1 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@366 -- # decimal 2 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@353 -- # local d=2 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@355 -- # echo 2 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@366 -- # ver2[v]=2 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@368 -- # return 0 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:17:31.339 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:31.339 --rc genhtml_branch_coverage=1 00:17:31.339 --rc genhtml_function_coverage=1 00:17:31.339 --rc genhtml_legend=1 00:17:31.339 --rc geninfo_all_blocks=1 00:17:31.339 --rc geninfo_unexecuted_blocks=1 00:17:31.339 00:17:31.339 ' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:17:31.339 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:31.339 --rc genhtml_branch_coverage=1 00:17:31.339 --rc genhtml_function_coverage=1 00:17:31.339 --rc genhtml_legend=1 00:17:31.339 --rc geninfo_all_blocks=1 00:17:31.339 --rc geninfo_unexecuted_blocks=1 00:17:31.339 00:17:31.339 ' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:17:31.339 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:31.339 --rc genhtml_branch_coverage=1 00:17:31.339 --rc genhtml_function_coverage=1 00:17:31.339 --rc genhtml_legend=1 00:17:31.339 --rc geninfo_all_blocks=1 00:17:31.339 --rc geninfo_unexecuted_blocks=1 00:17:31.339 00:17:31.339 ' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:17:31.339 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:31.339 --rc genhtml_branch_coverage=1 00:17:31.339 --rc genhtml_function_coverage=1 00:17:31.339 --rc genhtml_legend=1 00:17:31.339 --rc geninfo_all_blocks=1 00:17:31.339 --rc geninfo_unexecuted_blocks=1 00:17:31.339 00:17:31.339 ' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@7 -- # uname -s 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@15 -- # shopt -s extglob 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@5 -- # export PATH 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@51 -- # : 0 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:17:31.339 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@55 -- # have_pci_nics=0 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@12 -- # nvmftestinit 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@474 -- # prepare_net_devs 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@436 -- # local -g is_hw=no 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@438 -- # remove_spdk_ns 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@309 -- # xtrace_disable 00:17:31.339 08:53:21 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@315 -- # pci_devs=() 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@315 -- # local -a pci_devs 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@316 -- # pci_net_devs=() 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@317 -- # pci_drivers=() 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@317 -- # local -A pci_drivers 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@319 -- # net_devs=() 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@319 -- # local -ga net_devs 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@320 -- # e810=() 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@320 -- # local -ga e810 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@321 -- # x722=() 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@321 -- # local -ga x722 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@322 -- # mlx=() 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@322 -- # local -ga mlx 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:17:39.478 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:17:39.478 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:17:39.478 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@416 -- # [[ up == up ]] 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:17:39.479 Found net devices under 0000:4b:00.0: cvl_0_0 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@416 -- # [[ up == up ]] 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:17:39.479 Found net devices under 0000:4b:00.1: cvl_0_1 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@440 -- # is_hw=yes 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:17:39.479 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:17:39.479 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.690 ms 00:17:39.479 00:17:39.479 --- 10.0.0.2 ping statistics --- 00:17:39.479 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:39.479 rtt min/avg/max/mdev = 0.690/0.690/0.690/0.000 ms 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:17:39.479 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:17:39.479 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.315 ms 00:17:39.479 00:17:39.479 --- 10.0.0.1 ping statistics --- 00:17:39.479 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:39.479 rtt min/avg/max/mdev = 0.315/0.315/0.315/0.000 ms 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@448 -- # return 0 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@13 -- # nvmfappstart -m 0x2 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@507 -- # nvmfpid=1829771 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@508 -- # waitforlisten 1829771 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@831 -- # '[' -z 1829771 ']' 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:39.479 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:39.479 08:53:28 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:39.479 [2024-11-06 08:53:28.656669] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:17:39.479 [2024-11-06 08:53:28.656738] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:39.479 [2024-11-06 08:53:28.756074] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:39.479 [2024-11-06 08:53:28.806629] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:17:39.479 [2024-11-06 08:53:28.806681] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:17:39.479 [2024-11-06 08:53:28.806689] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:17:39.479 [2024-11-06 08:53:28.806697] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:17:39.479 [2024-11-06 08:53:28.806703] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:17:39.479 [2024-11-06 08:53:28.807453] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@864 -- # return 0 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:39.479 [2024-11-06 08:53:29.523410] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:39.479 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:39.480 [2024-11-06 08:53:29.539658] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:39.480 NULL1 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@19 -- # rpc_cmd bdev_wait_for_examine 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@561 -- # xtrace_disable 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:17:39.480 08:53:29 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/fused_ordering/fused_ordering -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:17:39.739 [2024-11-06 08:53:29.598203] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:17:39.739 [2024-11-06 08:53:29.598248] [ DPDK EAL parameters: fused_ordering --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1829802 ] 00:17:39.999 Attached to nqn.2016-06.io.spdk:cnode1 00:17:39.999 Namespace ID: 1 size: 1GB 00:17:39.999 fused_ordering(0) 00:17:39.999 fused_ordering(1) 00:17:39.999 fused_ordering(2) 00:17:39.999 fused_ordering(3) 00:17:39.999 fused_ordering(4) 00:17:39.999 fused_ordering(5) 00:17:39.999 fused_ordering(6) 00:17:39.999 fused_ordering(7) 00:17:39.999 fused_ordering(8) 00:17:39.999 fused_ordering(9) 00:17:39.999 fused_ordering(10) 00:17:39.999 fused_ordering(11) 00:17:39.999 fused_ordering(12) 00:17:39.999 fused_ordering(13) 00:17:39.999 fused_ordering(14) 00:17:39.999 fused_ordering(15) 00:17:39.999 fused_ordering(16) 00:17:39.999 fused_ordering(17) 00:17:39.999 fused_ordering(18) 00:17:39.999 fused_ordering(19) 00:17:39.999 fused_ordering(20) 00:17:39.999 fused_ordering(21) 00:17:39.999 fused_ordering(22) 00:17:39.999 fused_ordering(23) 00:17:39.999 fused_ordering(24) 00:17:39.999 fused_ordering(25) 00:17:39.999 fused_ordering(26) 00:17:39.999 fused_ordering(27) 00:17:39.999 fused_ordering(28) 00:17:39.999 fused_ordering(29) 00:17:39.999 fused_ordering(30) 00:17:39.999 fused_ordering(31) 00:17:39.999 fused_ordering(32) 00:17:39.999 fused_ordering(33) 00:17:39.999 fused_ordering(34) 00:17:39.999 fused_ordering(35) 00:17:39.999 fused_ordering(36) 00:17:39.999 fused_ordering(37) 00:17:39.999 fused_ordering(38) 00:17:39.999 fused_ordering(39) 00:17:39.999 fused_ordering(40) 00:17:39.999 fused_ordering(41) 00:17:39.999 fused_ordering(42) 00:17:39.999 fused_ordering(43) 00:17:39.999 fused_ordering(44) 00:17:39.999 fused_ordering(45) 00:17:39.999 fused_ordering(46) 00:17:39.999 fused_ordering(47) 00:17:39.999 fused_ordering(48) 00:17:39.999 fused_ordering(49) 00:17:39.999 fused_ordering(50) 00:17:39.999 fused_ordering(51) 00:17:39.999 fused_ordering(52) 00:17:39.999 fused_ordering(53) 00:17:39.999 fused_ordering(54) 00:17:39.999 fused_ordering(55) 00:17:39.999 fused_ordering(56) 00:17:39.999 fused_ordering(57) 00:17:39.999 fused_ordering(58) 00:17:39.999 fused_ordering(59) 00:17:39.999 fused_ordering(60) 00:17:39.999 fused_ordering(61) 00:17:39.999 fused_ordering(62) 00:17:39.999 fused_ordering(63) 00:17:39.999 fused_ordering(64) 00:17:39.999 fused_ordering(65) 00:17:39.999 fused_ordering(66) 00:17:39.999 fused_ordering(67) 00:17:39.999 fused_ordering(68) 00:17:39.999 fused_ordering(69) 00:17:39.999 fused_ordering(70) 00:17:39.999 fused_ordering(71) 00:17:39.999 fused_ordering(72) 00:17:39.999 fused_ordering(73) 00:17:39.999 fused_ordering(74) 00:17:39.999 fused_ordering(75) 00:17:39.999 fused_ordering(76) 00:17:39.999 fused_ordering(77) 00:17:39.999 fused_ordering(78) 00:17:39.999 fused_ordering(79) 00:17:39.999 fused_ordering(80) 00:17:39.999 fused_ordering(81) 00:17:39.999 fused_ordering(82) 00:17:39.999 fused_ordering(83) 00:17:39.999 fused_ordering(84) 00:17:39.999 fused_ordering(85) 00:17:39.999 fused_ordering(86) 00:17:39.999 fused_ordering(87) 00:17:39.999 fused_ordering(88) 00:17:39.999 fused_ordering(89) 00:17:39.999 fused_ordering(90) 00:17:39.999 fused_ordering(91) 00:17:39.999 fused_ordering(92) 00:17:39.999 fused_ordering(93) 00:17:39.999 fused_ordering(94) 00:17:39.999 fused_ordering(95) 00:17:39.999 fused_ordering(96) 00:17:39.999 fused_ordering(97) 00:17:39.999 fused_ordering(98) 00:17:39.999 fused_ordering(99) 00:17:39.999 fused_ordering(100) 00:17:39.999 fused_ordering(101) 00:17:39.999 fused_ordering(102) 00:17:39.999 fused_ordering(103) 00:17:39.999 fused_ordering(104) 00:17:39.999 fused_ordering(105) 00:17:39.999 fused_ordering(106) 00:17:39.999 fused_ordering(107) 00:17:39.999 fused_ordering(108) 00:17:39.999 fused_ordering(109) 00:17:39.999 fused_ordering(110) 00:17:39.999 fused_ordering(111) 00:17:39.999 fused_ordering(112) 00:17:39.999 fused_ordering(113) 00:17:39.999 fused_ordering(114) 00:17:39.999 fused_ordering(115) 00:17:39.999 fused_ordering(116) 00:17:39.999 fused_ordering(117) 00:17:39.999 fused_ordering(118) 00:17:39.999 fused_ordering(119) 00:17:39.999 fused_ordering(120) 00:17:39.999 fused_ordering(121) 00:17:39.999 fused_ordering(122) 00:17:39.999 fused_ordering(123) 00:17:39.999 fused_ordering(124) 00:17:39.999 fused_ordering(125) 00:17:39.999 fused_ordering(126) 00:17:39.999 fused_ordering(127) 00:17:39.999 fused_ordering(128) 00:17:39.999 fused_ordering(129) 00:17:39.999 fused_ordering(130) 00:17:39.999 fused_ordering(131) 00:17:39.999 fused_ordering(132) 00:17:39.999 fused_ordering(133) 00:17:39.999 fused_ordering(134) 00:17:39.999 fused_ordering(135) 00:17:39.999 fused_ordering(136) 00:17:39.999 fused_ordering(137) 00:17:39.999 fused_ordering(138) 00:17:39.999 fused_ordering(139) 00:17:39.999 fused_ordering(140) 00:17:39.999 fused_ordering(141) 00:17:39.999 fused_ordering(142) 00:17:39.999 fused_ordering(143) 00:17:39.999 fused_ordering(144) 00:17:39.999 fused_ordering(145) 00:17:39.999 fused_ordering(146) 00:17:39.999 fused_ordering(147) 00:17:39.999 fused_ordering(148) 00:17:39.999 fused_ordering(149) 00:17:39.999 fused_ordering(150) 00:17:39.999 fused_ordering(151) 00:17:39.999 fused_ordering(152) 00:17:39.999 fused_ordering(153) 00:17:39.999 fused_ordering(154) 00:17:39.999 fused_ordering(155) 00:17:39.999 fused_ordering(156) 00:17:39.999 fused_ordering(157) 00:17:39.999 fused_ordering(158) 00:17:39.999 fused_ordering(159) 00:17:39.999 fused_ordering(160) 00:17:40.000 fused_ordering(161) 00:17:40.000 fused_ordering(162) 00:17:40.000 fused_ordering(163) 00:17:40.000 fused_ordering(164) 00:17:40.000 fused_ordering(165) 00:17:40.000 fused_ordering(166) 00:17:40.000 fused_ordering(167) 00:17:40.000 fused_ordering(168) 00:17:40.000 fused_ordering(169) 00:17:40.000 fused_ordering(170) 00:17:40.000 fused_ordering(171) 00:17:40.000 fused_ordering(172) 00:17:40.000 fused_ordering(173) 00:17:40.000 fused_ordering(174) 00:17:40.000 fused_ordering(175) 00:17:40.000 fused_ordering(176) 00:17:40.000 fused_ordering(177) 00:17:40.000 fused_ordering(178) 00:17:40.000 fused_ordering(179) 00:17:40.000 fused_ordering(180) 00:17:40.000 fused_ordering(181) 00:17:40.000 fused_ordering(182) 00:17:40.000 fused_ordering(183) 00:17:40.000 fused_ordering(184) 00:17:40.000 fused_ordering(185) 00:17:40.000 fused_ordering(186) 00:17:40.000 fused_ordering(187) 00:17:40.000 fused_ordering(188) 00:17:40.000 fused_ordering(189) 00:17:40.000 fused_ordering(190) 00:17:40.000 fused_ordering(191) 00:17:40.000 fused_ordering(192) 00:17:40.000 fused_ordering(193) 00:17:40.000 fused_ordering(194) 00:17:40.000 fused_ordering(195) 00:17:40.000 fused_ordering(196) 00:17:40.000 fused_ordering(197) 00:17:40.000 fused_ordering(198) 00:17:40.000 fused_ordering(199) 00:17:40.000 fused_ordering(200) 00:17:40.000 fused_ordering(201) 00:17:40.000 fused_ordering(202) 00:17:40.000 fused_ordering(203) 00:17:40.000 fused_ordering(204) 00:17:40.000 fused_ordering(205) 00:17:40.260 fused_ordering(206) 00:17:40.260 fused_ordering(207) 00:17:40.260 fused_ordering(208) 00:17:40.260 fused_ordering(209) 00:17:40.260 fused_ordering(210) 00:17:40.260 fused_ordering(211) 00:17:40.260 fused_ordering(212) 00:17:40.260 fused_ordering(213) 00:17:40.260 fused_ordering(214) 00:17:40.260 fused_ordering(215) 00:17:40.260 fused_ordering(216) 00:17:40.260 fused_ordering(217) 00:17:40.260 fused_ordering(218) 00:17:40.260 fused_ordering(219) 00:17:40.260 fused_ordering(220) 00:17:40.260 fused_ordering(221) 00:17:40.260 fused_ordering(222) 00:17:40.260 fused_ordering(223) 00:17:40.260 fused_ordering(224) 00:17:40.260 fused_ordering(225) 00:17:40.260 fused_ordering(226) 00:17:40.260 fused_ordering(227) 00:17:40.260 fused_ordering(228) 00:17:40.260 fused_ordering(229) 00:17:40.260 fused_ordering(230) 00:17:40.260 fused_ordering(231) 00:17:40.260 fused_ordering(232) 00:17:40.260 fused_ordering(233) 00:17:40.260 fused_ordering(234) 00:17:40.260 fused_ordering(235) 00:17:40.260 fused_ordering(236) 00:17:40.260 fused_ordering(237) 00:17:40.260 fused_ordering(238) 00:17:40.260 fused_ordering(239) 00:17:40.260 fused_ordering(240) 00:17:40.260 fused_ordering(241) 00:17:40.260 fused_ordering(242) 00:17:40.260 fused_ordering(243) 00:17:40.260 fused_ordering(244) 00:17:40.260 fused_ordering(245) 00:17:40.260 fused_ordering(246) 00:17:40.260 fused_ordering(247) 00:17:40.260 fused_ordering(248) 00:17:40.260 fused_ordering(249) 00:17:40.260 fused_ordering(250) 00:17:40.260 fused_ordering(251) 00:17:40.260 fused_ordering(252) 00:17:40.260 fused_ordering(253) 00:17:40.260 fused_ordering(254) 00:17:40.260 fused_ordering(255) 00:17:40.260 fused_ordering(256) 00:17:40.260 fused_ordering(257) 00:17:40.260 fused_ordering(258) 00:17:40.260 fused_ordering(259) 00:17:40.260 fused_ordering(260) 00:17:40.260 fused_ordering(261) 00:17:40.260 fused_ordering(262) 00:17:40.260 fused_ordering(263) 00:17:40.260 fused_ordering(264) 00:17:40.260 fused_ordering(265) 00:17:40.260 fused_ordering(266) 00:17:40.260 fused_ordering(267) 00:17:40.260 fused_ordering(268) 00:17:40.260 fused_ordering(269) 00:17:40.260 fused_ordering(270) 00:17:40.260 fused_ordering(271) 00:17:40.260 fused_ordering(272) 00:17:40.260 fused_ordering(273) 00:17:40.260 fused_ordering(274) 00:17:40.260 fused_ordering(275) 00:17:40.260 fused_ordering(276) 00:17:40.260 fused_ordering(277) 00:17:40.260 fused_ordering(278) 00:17:40.260 fused_ordering(279) 00:17:40.260 fused_ordering(280) 00:17:40.260 fused_ordering(281) 00:17:40.260 fused_ordering(282) 00:17:40.260 fused_ordering(283) 00:17:40.260 fused_ordering(284) 00:17:40.260 fused_ordering(285) 00:17:40.260 fused_ordering(286) 00:17:40.260 fused_ordering(287) 00:17:40.260 fused_ordering(288) 00:17:40.260 fused_ordering(289) 00:17:40.260 fused_ordering(290) 00:17:40.260 fused_ordering(291) 00:17:40.260 fused_ordering(292) 00:17:40.260 fused_ordering(293) 00:17:40.260 fused_ordering(294) 00:17:40.260 fused_ordering(295) 00:17:40.260 fused_ordering(296) 00:17:40.260 fused_ordering(297) 00:17:40.260 fused_ordering(298) 00:17:40.260 fused_ordering(299) 00:17:40.260 fused_ordering(300) 00:17:40.260 fused_ordering(301) 00:17:40.260 fused_ordering(302) 00:17:40.260 fused_ordering(303) 00:17:40.260 fused_ordering(304) 00:17:40.260 fused_ordering(305) 00:17:40.260 fused_ordering(306) 00:17:40.260 fused_ordering(307) 00:17:40.260 fused_ordering(308) 00:17:40.260 fused_ordering(309) 00:17:40.260 fused_ordering(310) 00:17:40.260 fused_ordering(311) 00:17:40.260 fused_ordering(312) 00:17:40.260 fused_ordering(313) 00:17:40.260 fused_ordering(314) 00:17:40.260 fused_ordering(315) 00:17:40.260 fused_ordering(316) 00:17:40.260 fused_ordering(317) 00:17:40.260 fused_ordering(318) 00:17:40.260 fused_ordering(319) 00:17:40.260 fused_ordering(320) 00:17:40.260 fused_ordering(321) 00:17:40.260 fused_ordering(322) 00:17:40.260 fused_ordering(323) 00:17:40.260 fused_ordering(324) 00:17:40.260 fused_ordering(325) 00:17:40.260 fused_ordering(326) 00:17:40.260 fused_ordering(327) 00:17:40.260 fused_ordering(328) 00:17:40.260 fused_ordering(329) 00:17:40.260 fused_ordering(330) 00:17:40.260 fused_ordering(331) 00:17:40.260 fused_ordering(332) 00:17:40.260 fused_ordering(333) 00:17:40.260 fused_ordering(334) 00:17:40.260 fused_ordering(335) 00:17:40.261 fused_ordering(336) 00:17:40.261 fused_ordering(337) 00:17:40.261 fused_ordering(338) 00:17:40.261 fused_ordering(339) 00:17:40.261 fused_ordering(340) 00:17:40.261 fused_ordering(341) 00:17:40.261 fused_ordering(342) 00:17:40.261 fused_ordering(343) 00:17:40.261 fused_ordering(344) 00:17:40.261 fused_ordering(345) 00:17:40.261 fused_ordering(346) 00:17:40.261 fused_ordering(347) 00:17:40.261 fused_ordering(348) 00:17:40.261 fused_ordering(349) 00:17:40.261 fused_ordering(350) 00:17:40.261 fused_ordering(351) 00:17:40.261 fused_ordering(352) 00:17:40.261 fused_ordering(353) 00:17:40.261 fused_ordering(354) 00:17:40.261 fused_ordering(355) 00:17:40.261 fused_ordering(356) 00:17:40.261 fused_ordering(357) 00:17:40.261 fused_ordering(358) 00:17:40.261 fused_ordering(359) 00:17:40.261 fused_ordering(360) 00:17:40.261 fused_ordering(361) 00:17:40.261 fused_ordering(362) 00:17:40.261 fused_ordering(363) 00:17:40.261 fused_ordering(364) 00:17:40.261 fused_ordering(365) 00:17:40.261 fused_ordering(366) 00:17:40.261 fused_ordering(367) 00:17:40.261 fused_ordering(368) 00:17:40.261 fused_ordering(369) 00:17:40.261 fused_ordering(370) 00:17:40.261 fused_ordering(371) 00:17:40.261 fused_ordering(372) 00:17:40.261 fused_ordering(373) 00:17:40.261 fused_ordering(374) 00:17:40.261 fused_ordering(375) 00:17:40.261 fused_ordering(376) 00:17:40.261 fused_ordering(377) 00:17:40.261 fused_ordering(378) 00:17:40.261 fused_ordering(379) 00:17:40.261 fused_ordering(380) 00:17:40.261 fused_ordering(381) 00:17:40.261 fused_ordering(382) 00:17:40.261 fused_ordering(383) 00:17:40.261 fused_ordering(384) 00:17:40.261 fused_ordering(385) 00:17:40.261 fused_ordering(386) 00:17:40.261 fused_ordering(387) 00:17:40.261 fused_ordering(388) 00:17:40.261 fused_ordering(389) 00:17:40.261 fused_ordering(390) 00:17:40.261 fused_ordering(391) 00:17:40.261 fused_ordering(392) 00:17:40.261 fused_ordering(393) 00:17:40.261 fused_ordering(394) 00:17:40.261 fused_ordering(395) 00:17:40.261 fused_ordering(396) 00:17:40.261 fused_ordering(397) 00:17:40.261 fused_ordering(398) 00:17:40.261 fused_ordering(399) 00:17:40.261 fused_ordering(400) 00:17:40.261 fused_ordering(401) 00:17:40.261 fused_ordering(402) 00:17:40.261 fused_ordering(403) 00:17:40.261 fused_ordering(404) 00:17:40.261 fused_ordering(405) 00:17:40.261 fused_ordering(406) 00:17:40.261 fused_ordering(407) 00:17:40.261 fused_ordering(408) 00:17:40.261 fused_ordering(409) 00:17:40.261 fused_ordering(410) 00:17:40.911 fused_ordering(411) 00:17:40.911 fused_ordering(412) 00:17:40.911 fused_ordering(413) 00:17:40.911 fused_ordering(414) 00:17:40.911 fused_ordering(415) 00:17:40.911 fused_ordering(416) 00:17:40.911 fused_ordering(417) 00:17:40.911 fused_ordering(418) 00:17:40.911 fused_ordering(419) 00:17:40.911 fused_ordering(420) 00:17:40.911 fused_ordering(421) 00:17:40.911 fused_ordering(422) 00:17:40.911 fused_ordering(423) 00:17:40.911 fused_ordering(424) 00:17:40.911 fused_ordering(425) 00:17:40.911 fused_ordering(426) 00:17:40.911 fused_ordering(427) 00:17:40.911 fused_ordering(428) 00:17:40.911 fused_ordering(429) 00:17:40.911 fused_ordering(430) 00:17:40.911 fused_ordering(431) 00:17:40.911 fused_ordering(432) 00:17:40.911 fused_ordering(433) 00:17:40.911 fused_ordering(434) 00:17:40.911 fused_ordering(435) 00:17:40.911 fused_ordering(436) 00:17:40.911 fused_ordering(437) 00:17:40.911 fused_ordering(438) 00:17:40.911 fused_ordering(439) 00:17:40.911 fused_ordering(440) 00:17:40.911 fused_ordering(441) 00:17:40.911 fused_ordering(442) 00:17:40.911 fused_ordering(443) 00:17:40.911 fused_ordering(444) 00:17:40.911 fused_ordering(445) 00:17:40.911 fused_ordering(446) 00:17:40.911 fused_ordering(447) 00:17:40.911 fused_ordering(448) 00:17:40.911 fused_ordering(449) 00:17:40.911 fused_ordering(450) 00:17:40.911 fused_ordering(451) 00:17:40.911 fused_ordering(452) 00:17:40.911 fused_ordering(453) 00:17:40.911 fused_ordering(454) 00:17:40.911 fused_ordering(455) 00:17:40.911 fused_ordering(456) 00:17:40.911 fused_ordering(457) 00:17:40.911 fused_ordering(458) 00:17:40.911 fused_ordering(459) 00:17:40.911 fused_ordering(460) 00:17:40.911 fused_ordering(461) 00:17:40.911 fused_ordering(462) 00:17:40.911 fused_ordering(463) 00:17:40.911 fused_ordering(464) 00:17:40.911 fused_ordering(465) 00:17:40.911 fused_ordering(466) 00:17:40.911 fused_ordering(467) 00:17:40.911 fused_ordering(468) 00:17:40.911 fused_ordering(469) 00:17:40.911 fused_ordering(470) 00:17:40.911 fused_ordering(471) 00:17:40.911 fused_ordering(472) 00:17:40.911 fused_ordering(473) 00:17:40.911 fused_ordering(474) 00:17:40.911 fused_ordering(475) 00:17:40.911 fused_ordering(476) 00:17:40.911 fused_ordering(477) 00:17:40.911 fused_ordering(478) 00:17:40.911 fused_ordering(479) 00:17:40.911 fused_ordering(480) 00:17:40.911 fused_ordering(481) 00:17:40.911 fused_ordering(482) 00:17:40.911 fused_ordering(483) 00:17:40.911 fused_ordering(484) 00:17:40.911 fused_ordering(485) 00:17:40.911 fused_ordering(486) 00:17:40.911 fused_ordering(487) 00:17:40.911 fused_ordering(488) 00:17:40.911 fused_ordering(489) 00:17:40.911 fused_ordering(490) 00:17:40.911 fused_ordering(491) 00:17:40.911 fused_ordering(492) 00:17:40.911 fused_ordering(493) 00:17:40.911 fused_ordering(494) 00:17:40.911 fused_ordering(495) 00:17:40.911 fused_ordering(496) 00:17:40.911 fused_ordering(497) 00:17:40.911 fused_ordering(498) 00:17:40.911 fused_ordering(499) 00:17:40.911 fused_ordering(500) 00:17:40.911 fused_ordering(501) 00:17:40.911 fused_ordering(502) 00:17:40.911 fused_ordering(503) 00:17:40.911 fused_ordering(504) 00:17:40.911 fused_ordering(505) 00:17:40.911 fused_ordering(506) 00:17:40.911 fused_ordering(507) 00:17:40.911 fused_ordering(508) 00:17:40.911 fused_ordering(509) 00:17:40.911 fused_ordering(510) 00:17:40.911 fused_ordering(511) 00:17:40.911 fused_ordering(512) 00:17:40.911 fused_ordering(513) 00:17:40.911 fused_ordering(514) 00:17:40.911 fused_ordering(515) 00:17:40.911 fused_ordering(516) 00:17:40.911 fused_ordering(517) 00:17:40.911 fused_ordering(518) 00:17:40.911 fused_ordering(519) 00:17:40.911 fused_ordering(520) 00:17:40.911 fused_ordering(521) 00:17:40.911 fused_ordering(522) 00:17:40.911 fused_ordering(523) 00:17:40.911 fused_ordering(524) 00:17:40.911 fused_ordering(525) 00:17:40.911 fused_ordering(526) 00:17:40.911 fused_ordering(527) 00:17:40.911 fused_ordering(528) 00:17:40.911 fused_ordering(529) 00:17:40.911 fused_ordering(530) 00:17:40.911 fused_ordering(531) 00:17:40.911 fused_ordering(532) 00:17:40.911 fused_ordering(533) 00:17:40.911 fused_ordering(534) 00:17:40.911 fused_ordering(535) 00:17:40.911 fused_ordering(536) 00:17:40.911 fused_ordering(537) 00:17:40.911 fused_ordering(538) 00:17:40.911 fused_ordering(539) 00:17:40.911 fused_ordering(540) 00:17:40.911 fused_ordering(541) 00:17:40.911 fused_ordering(542) 00:17:40.911 fused_ordering(543) 00:17:40.911 fused_ordering(544) 00:17:40.911 fused_ordering(545) 00:17:40.911 fused_ordering(546) 00:17:40.911 fused_ordering(547) 00:17:40.911 fused_ordering(548) 00:17:40.911 fused_ordering(549) 00:17:40.911 fused_ordering(550) 00:17:40.911 fused_ordering(551) 00:17:40.911 fused_ordering(552) 00:17:40.911 fused_ordering(553) 00:17:40.911 fused_ordering(554) 00:17:40.911 fused_ordering(555) 00:17:40.911 fused_ordering(556) 00:17:40.911 fused_ordering(557) 00:17:40.911 fused_ordering(558) 00:17:40.911 fused_ordering(559) 00:17:40.911 fused_ordering(560) 00:17:40.911 fused_ordering(561) 00:17:40.911 fused_ordering(562) 00:17:40.911 fused_ordering(563) 00:17:40.911 fused_ordering(564) 00:17:40.911 fused_ordering(565) 00:17:40.911 fused_ordering(566) 00:17:40.911 fused_ordering(567) 00:17:40.911 fused_ordering(568) 00:17:40.911 fused_ordering(569) 00:17:40.911 fused_ordering(570) 00:17:40.911 fused_ordering(571) 00:17:40.911 fused_ordering(572) 00:17:40.911 fused_ordering(573) 00:17:40.911 fused_ordering(574) 00:17:40.911 fused_ordering(575) 00:17:40.911 fused_ordering(576) 00:17:40.911 fused_ordering(577) 00:17:40.911 fused_ordering(578) 00:17:40.911 fused_ordering(579) 00:17:40.911 fused_ordering(580) 00:17:40.911 fused_ordering(581) 00:17:40.911 fused_ordering(582) 00:17:40.911 fused_ordering(583) 00:17:40.911 fused_ordering(584) 00:17:40.911 fused_ordering(585) 00:17:40.911 fused_ordering(586) 00:17:40.911 fused_ordering(587) 00:17:40.911 fused_ordering(588) 00:17:40.911 fused_ordering(589) 00:17:40.911 fused_ordering(590) 00:17:40.911 fused_ordering(591) 00:17:40.911 fused_ordering(592) 00:17:40.911 fused_ordering(593) 00:17:40.911 fused_ordering(594) 00:17:40.911 fused_ordering(595) 00:17:40.911 fused_ordering(596) 00:17:40.911 fused_ordering(597) 00:17:40.911 fused_ordering(598) 00:17:40.911 fused_ordering(599) 00:17:40.911 fused_ordering(600) 00:17:40.911 fused_ordering(601) 00:17:40.911 fused_ordering(602) 00:17:40.911 fused_ordering(603) 00:17:40.911 fused_ordering(604) 00:17:40.911 fused_ordering(605) 00:17:40.911 fused_ordering(606) 00:17:40.911 fused_ordering(607) 00:17:40.911 fused_ordering(608) 00:17:40.911 fused_ordering(609) 00:17:40.911 fused_ordering(610) 00:17:40.911 fused_ordering(611) 00:17:40.911 fused_ordering(612) 00:17:40.911 fused_ordering(613) 00:17:40.911 fused_ordering(614) 00:17:40.911 fused_ordering(615) 00:17:41.172 fused_ordering(616) 00:17:41.172 fused_ordering(617) 00:17:41.172 fused_ordering(618) 00:17:41.172 fused_ordering(619) 00:17:41.172 fused_ordering(620) 00:17:41.172 fused_ordering(621) 00:17:41.172 fused_ordering(622) 00:17:41.172 fused_ordering(623) 00:17:41.172 fused_ordering(624) 00:17:41.172 fused_ordering(625) 00:17:41.172 fused_ordering(626) 00:17:41.172 fused_ordering(627) 00:17:41.172 fused_ordering(628) 00:17:41.172 fused_ordering(629) 00:17:41.172 fused_ordering(630) 00:17:41.172 fused_ordering(631) 00:17:41.172 fused_ordering(632) 00:17:41.172 fused_ordering(633) 00:17:41.172 fused_ordering(634) 00:17:41.172 fused_ordering(635) 00:17:41.172 fused_ordering(636) 00:17:41.172 fused_ordering(637) 00:17:41.172 fused_ordering(638) 00:17:41.172 fused_ordering(639) 00:17:41.172 fused_ordering(640) 00:17:41.172 fused_ordering(641) 00:17:41.172 fused_ordering(642) 00:17:41.172 fused_ordering(643) 00:17:41.172 fused_ordering(644) 00:17:41.172 fused_ordering(645) 00:17:41.172 fused_ordering(646) 00:17:41.172 fused_ordering(647) 00:17:41.172 fused_ordering(648) 00:17:41.172 fused_ordering(649) 00:17:41.172 fused_ordering(650) 00:17:41.172 fused_ordering(651) 00:17:41.172 fused_ordering(652) 00:17:41.172 fused_ordering(653) 00:17:41.172 fused_ordering(654) 00:17:41.172 fused_ordering(655) 00:17:41.172 fused_ordering(656) 00:17:41.172 fused_ordering(657) 00:17:41.172 fused_ordering(658) 00:17:41.172 fused_ordering(659) 00:17:41.172 fused_ordering(660) 00:17:41.172 fused_ordering(661) 00:17:41.172 fused_ordering(662) 00:17:41.172 fused_ordering(663) 00:17:41.172 fused_ordering(664) 00:17:41.172 fused_ordering(665) 00:17:41.172 fused_ordering(666) 00:17:41.172 fused_ordering(667) 00:17:41.172 fused_ordering(668) 00:17:41.172 fused_ordering(669) 00:17:41.172 fused_ordering(670) 00:17:41.172 fused_ordering(671) 00:17:41.172 fused_ordering(672) 00:17:41.172 fused_ordering(673) 00:17:41.172 fused_ordering(674) 00:17:41.172 fused_ordering(675) 00:17:41.172 fused_ordering(676) 00:17:41.172 fused_ordering(677) 00:17:41.172 fused_ordering(678) 00:17:41.172 fused_ordering(679) 00:17:41.172 fused_ordering(680) 00:17:41.172 fused_ordering(681) 00:17:41.172 fused_ordering(682) 00:17:41.172 fused_ordering(683) 00:17:41.172 fused_ordering(684) 00:17:41.172 fused_ordering(685) 00:17:41.172 fused_ordering(686) 00:17:41.172 fused_ordering(687) 00:17:41.172 fused_ordering(688) 00:17:41.172 fused_ordering(689) 00:17:41.172 fused_ordering(690) 00:17:41.172 fused_ordering(691) 00:17:41.172 fused_ordering(692) 00:17:41.172 fused_ordering(693) 00:17:41.172 fused_ordering(694) 00:17:41.172 fused_ordering(695) 00:17:41.172 fused_ordering(696) 00:17:41.172 fused_ordering(697) 00:17:41.172 fused_ordering(698) 00:17:41.172 fused_ordering(699) 00:17:41.172 fused_ordering(700) 00:17:41.172 fused_ordering(701) 00:17:41.172 fused_ordering(702) 00:17:41.172 fused_ordering(703) 00:17:41.172 fused_ordering(704) 00:17:41.172 fused_ordering(705) 00:17:41.172 fused_ordering(706) 00:17:41.172 fused_ordering(707) 00:17:41.172 fused_ordering(708) 00:17:41.172 fused_ordering(709) 00:17:41.172 fused_ordering(710) 00:17:41.172 fused_ordering(711) 00:17:41.172 fused_ordering(712) 00:17:41.172 fused_ordering(713) 00:17:41.172 fused_ordering(714) 00:17:41.172 fused_ordering(715) 00:17:41.172 fused_ordering(716) 00:17:41.172 fused_ordering(717) 00:17:41.172 fused_ordering(718) 00:17:41.172 fused_ordering(719) 00:17:41.172 fused_ordering(720) 00:17:41.172 fused_ordering(721) 00:17:41.172 fused_ordering(722) 00:17:41.172 fused_ordering(723) 00:17:41.172 fused_ordering(724) 00:17:41.172 fused_ordering(725) 00:17:41.172 fused_ordering(726) 00:17:41.172 fused_ordering(727) 00:17:41.172 fused_ordering(728) 00:17:41.172 fused_ordering(729) 00:17:41.172 fused_ordering(730) 00:17:41.172 fused_ordering(731) 00:17:41.172 fused_ordering(732) 00:17:41.172 fused_ordering(733) 00:17:41.172 fused_ordering(734) 00:17:41.172 fused_ordering(735) 00:17:41.172 fused_ordering(736) 00:17:41.172 fused_ordering(737) 00:17:41.172 fused_ordering(738) 00:17:41.172 fused_ordering(739) 00:17:41.172 fused_ordering(740) 00:17:41.172 fused_ordering(741) 00:17:41.172 fused_ordering(742) 00:17:41.172 fused_ordering(743) 00:17:41.172 fused_ordering(744) 00:17:41.172 fused_ordering(745) 00:17:41.172 fused_ordering(746) 00:17:41.172 fused_ordering(747) 00:17:41.172 fused_ordering(748) 00:17:41.172 fused_ordering(749) 00:17:41.172 fused_ordering(750) 00:17:41.172 fused_ordering(751) 00:17:41.172 fused_ordering(752) 00:17:41.172 fused_ordering(753) 00:17:41.172 fused_ordering(754) 00:17:41.172 fused_ordering(755) 00:17:41.172 fused_ordering(756) 00:17:41.172 fused_ordering(757) 00:17:41.172 fused_ordering(758) 00:17:41.172 fused_ordering(759) 00:17:41.172 fused_ordering(760) 00:17:41.172 fused_ordering(761) 00:17:41.172 fused_ordering(762) 00:17:41.172 fused_ordering(763) 00:17:41.172 fused_ordering(764) 00:17:41.172 fused_ordering(765) 00:17:41.172 fused_ordering(766) 00:17:41.172 fused_ordering(767) 00:17:41.172 fused_ordering(768) 00:17:41.172 fused_ordering(769) 00:17:41.172 fused_ordering(770) 00:17:41.172 fused_ordering(771) 00:17:41.172 fused_ordering(772) 00:17:41.172 fused_ordering(773) 00:17:41.172 fused_ordering(774) 00:17:41.172 fused_ordering(775) 00:17:41.172 fused_ordering(776) 00:17:41.172 fused_ordering(777) 00:17:41.172 fused_ordering(778) 00:17:41.172 fused_ordering(779) 00:17:41.172 fused_ordering(780) 00:17:41.172 fused_ordering(781) 00:17:41.172 fused_ordering(782) 00:17:41.172 fused_ordering(783) 00:17:41.172 fused_ordering(784) 00:17:41.172 fused_ordering(785) 00:17:41.172 fused_ordering(786) 00:17:41.172 fused_ordering(787) 00:17:41.172 fused_ordering(788) 00:17:41.172 fused_ordering(789) 00:17:41.172 fused_ordering(790) 00:17:41.172 fused_ordering(791) 00:17:41.172 fused_ordering(792) 00:17:41.172 fused_ordering(793) 00:17:41.173 fused_ordering(794) 00:17:41.173 fused_ordering(795) 00:17:41.173 fused_ordering(796) 00:17:41.173 fused_ordering(797) 00:17:41.173 fused_ordering(798) 00:17:41.173 fused_ordering(799) 00:17:41.173 fused_ordering(800) 00:17:41.173 fused_ordering(801) 00:17:41.173 fused_ordering(802) 00:17:41.173 fused_ordering(803) 00:17:41.173 fused_ordering(804) 00:17:41.173 fused_ordering(805) 00:17:41.173 fused_ordering(806) 00:17:41.173 fused_ordering(807) 00:17:41.173 fused_ordering(808) 00:17:41.173 fused_ordering(809) 00:17:41.173 fused_ordering(810) 00:17:41.173 fused_ordering(811) 00:17:41.173 fused_ordering(812) 00:17:41.173 fused_ordering(813) 00:17:41.173 fused_ordering(814) 00:17:41.173 fused_ordering(815) 00:17:41.173 fused_ordering(816) 00:17:41.173 fused_ordering(817) 00:17:41.173 fused_ordering(818) 00:17:41.173 fused_ordering(819) 00:17:41.173 fused_ordering(820) 00:17:41.745 fused_ordering(821) 00:17:41.745 fused_ordering(822) 00:17:41.745 fused_ordering(823) 00:17:41.745 fused_ordering(824) 00:17:41.745 fused_ordering(825) 00:17:41.745 fused_ordering(826) 00:17:41.745 fused_ordering(827) 00:17:41.745 fused_ordering(828) 00:17:41.745 fused_ordering(829) 00:17:41.745 fused_ordering(830) 00:17:41.745 fused_ordering(831) 00:17:41.745 fused_ordering(832) 00:17:41.745 fused_ordering(833) 00:17:41.745 fused_ordering(834) 00:17:41.745 fused_ordering(835) 00:17:41.745 fused_ordering(836) 00:17:41.745 fused_ordering(837) 00:17:41.745 fused_ordering(838) 00:17:41.745 fused_ordering(839) 00:17:41.745 fused_ordering(840) 00:17:41.745 fused_ordering(841) 00:17:41.745 fused_ordering(842) 00:17:41.745 fused_ordering(843) 00:17:41.745 fused_ordering(844) 00:17:41.745 fused_ordering(845) 00:17:41.745 fused_ordering(846) 00:17:41.745 fused_ordering(847) 00:17:41.745 fused_ordering(848) 00:17:41.745 fused_ordering(849) 00:17:41.745 fused_ordering(850) 00:17:41.745 fused_ordering(851) 00:17:41.745 fused_ordering(852) 00:17:41.745 fused_ordering(853) 00:17:41.745 fused_ordering(854) 00:17:41.745 fused_ordering(855) 00:17:41.745 fused_ordering(856) 00:17:41.745 fused_ordering(857) 00:17:41.745 fused_ordering(858) 00:17:41.745 fused_ordering(859) 00:17:41.745 fused_ordering(860) 00:17:41.745 fused_ordering(861) 00:17:41.745 fused_ordering(862) 00:17:41.745 fused_ordering(863) 00:17:41.745 fused_ordering(864) 00:17:41.745 fused_ordering(865) 00:17:41.745 fused_ordering(866) 00:17:41.745 fused_ordering(867) 00:17:41.745 fused_ordering(868) 00:17:41.745 fused_ordering(869) 00:17:41.745 fused_ordering(870) 00:17:41.745 fused_ordering(871) 00:17:41.745 fused_ordering(872) 00:17:41.745 fused_ordering(873) 00:17:41.745 fused_ordering(874) 00:17:41.745 fused_ordering(875) 00:17:41.745 fused_ordering(876) 00:17:41.745 fused_ordering(877) 00:17:41.745 fused_ordering(878) 00:17:41.745 fused_ordering(879) 00:17:41.745 fused_ordering(880) 00:17:41.745 fused_ordering(881) 00:17:41.745 fused_ordering(882) 00:17:41.745 fused_ordering(883) 00:17:41.745 fused_ordering(884) 00:17:41.745 fused_ordering(885) 00:17:41.745 fused_ordering(886) 00:17:41.745 fused_ordering(887) 00:17:41.745 fused_ordering(888) 00:17:41.745 fused_ordering(889) 00:17:41.745 fused_ordering(890) 00:17:41.745 fused_ordering(891) 00:17:41.745 fused_ordering(892) 00:17:41.745 fused_ordering(893) 00:17:41.745 fused_ordering(894) 00:17:41.745 fused_ordering(895) 00:17:41.745 fused_ordering(896) 00:17:41.745 fused_ordering(897) 00:17:41.745 fused_ordering(898) 00:17:41.745 fused_ordering(899) 00:17:41.745 fused_ordering(900) 00:17:41.745 fused_ordering(901) 00:17:41.745 fused_ordering(902) 00:17:41.745 fused_ordering(903) 00:17:41.745 fused_ordering(904) 00:17:41.745 fused_ordering(905) 00:17:41.745 fused_ordering(906) 00:17:41.745 fused_ordering(907) 00:17:41.745 fused_ordering(908) 00:17:41.745 fused_ordering(909) 00:17:41.745 fused_ordering(910) 00:17:41.745 fused_ordering(911) 00:17:41.745 fused_ordering(912) 00:17:41.745 fused_ordering(913) 00:17:41.745 fused_ordering(914) 00:17:41.745 fused_ordering(915) 00:17:41.745 fused_ordering(916) 00:17:41.745 fused_ordering(917) 00:17:41.745 fused_ordering(918) 00:17:41.745 fused_ordering(919) 00:17:41.745 fused_ordering(920) 00:17:41.745 fused_ordering(921) 00:17:41.745 fused_ordering(922) 00:17:41.745 fused_ordering(923) 00:17:41.745 fused_ordering(924) 00:17:41.745 fused_ordering(925) 00:17:41.745 fused_ordering(926) 00:17:41.745 fused_ordering(927) 00:17:41.745 fused_ordering(928) 00:17:41.745 fused_ordering(929) 00:17:41.745 fused_ordering(930) 00:17:41.745 fused_ordering(931) 00:17:41.745 fused_ordering(932) 00:17:41.746 fused_ordering(933) 00:17:41.746 fused_ordering(934) 00:17:41.746 fused_ordering(935) 00:17:41.746 fused_ordering(936) 00:17:41.746 fused_ordering(937) 00:17:41.746 fused_ordering(938) 00:17:41.746 fused_ordering(939) 00:17:41.746 fused_ordering(940) 00:17:41.746 fused_ordering(941) 00:17:41.746 fused_ordering(942) 00:17:41.746 fused_ordering(943) 00:17:41.746 fused_ordering(944) 00:17:41.746 fused_ordering(945) 00:17:41.746 fused_ordering(946) 00:17:41.746 fused_ordering(947) 00:17:41.746 fused_ordering(948) 00:17:41.746 fused_ordering(949) 00:17:41.746 fused_ordering(950) 00:17:41.746 fused_ordering(951) 00:17:41.746 fused_ordering(952) 00:17:41.746 fused_ordering(953) 00:17:41.746 fused_ordering(954) 00:17:41.746 fused_ordering(955) 00:17:41.746 fused_ordering(956) 00:17:41.746 fused_ordering(957) 00:17:41.746 fused_ordering(958) 00:17:41.746 fused_ordering(959) 00:17:41.746 fused_ordering(960) 00:17:41.746 fused_ordering(961) 00:17:41.746 fused_ordering(962) 00:17:41.746 fused_ordering(963) 00:17:41.746 fused_ordering(964) 00:17:41.746 fused_ordering(965) 00:17:41.746 fused_ordering(966) 00:17:41.746 fused_ordering(967) 00:17:41.746 fused_ordering(968) 00:17:41.746 fused_ordering(969) 00:17:41.746 fused_ordering(970) 00:17:41.746 fused_ordering(971) 00:17:41.746 fused_ordering(972) 00:17:41.746 fused_ordering(973) 00:17:41.746 fused_ordering(974) 00:17:41.746 fused_ordering(975) 00:17:41.746 fused_ordering(976) 00:17:41.746 fused_ordering(977) 00:17:41.746 fused_ordering(978) 00:17:41.746 fused_ordering(979) 00:17:41.746 fused_ordering(980) 00:17:41.746 fused_ordering(981) 00:17:41.746 fused_ordering(982) 00:17:41.746 fused_ordering(983) 00:17:41.746 fused_ordering(984) 00:17:41.746 fused_ordering(985) 00:17:41.746 fused_ordering(986) 00:17:41.746 fused_ordering(987) 00:17:41.746 fused_ordering(988) 00:17:41.746 fused_ordering(989) 00:17:41.746 fused_ordering(990) 00:17:41.746 fused_ordering(991) 00:17:41.746 fused_ordering(992) 00:17:41.746 fused_ordering(993) 00:17:41.746 fused_ordering(994) 00:17:41.746 fused_ordering(995) 00:17:41.746 fused_ordering(996) 00:17:41.746 fused_ordering(997) 00:17:41.746 fused_ordering(998) 00:17:41.746 fused_ordering(999) 00:17:41.746 fused_ordering(1000) 00:17:41.746 fused_ordering(1001) 00:17:41.746 fused_ordering(1002) 00:17:41.746 fused_ordering(1003) 00:17:41.746 fused_ordering(1004) 00:17:41.746 fused_ordering(1005) 00:17:41.746 fused_ordering(1006) 00:17:41.746 fused_ordering(1007) 00:17:41.746 fused_ordering(1008) 00:17:41.746 fused_ordering(1009) 00:17:41.746 fused_ordering(1010) 00:17:41.746 fused_ordering(1011) 00:17:41.746 fused_ordering(1012) 00:17:41.746 fused_ordering(1013) 00:17:41.746 fused_ordering(1014) 00:17:41.746 fused_ordering(1015) 00:17:41.746 fused_ordering(1016) 00:17:41.746 fused_ordering(1017) 00:17:41.746 fused_ordering(1018) 00:17:41.746 fused_ordering(1019) 00:17:41.746 fused_ordering(1020) 00:17:41.746 fused_ordering(1021) 00:17:41.746 fused_ordering(1022) 00:17:41.746 fused_ordering(1023) 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@23 -- # trap - SIGINT SIGTERM EXIT 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- target/fused_ordering.sh@25 -- # nvmftestfini 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@514 -- # nvmfcleanup 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@121 -- # sync 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@124 -- # set +e 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@125 -- # for i in {1..20} 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:17:41.746 rmmod nvme_tcp 00:17:41.746 rmmod nvme_fabrics 00:17:41.746 rmmod nvme_keyring 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@128 -- # set -e 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@129 -- # return 0 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@515 -- # '[' -n 1829771 ']' 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@516 -- # killprocess 1829771 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@950 -- # '[' -z 1829771 ']' 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@954 -- # kill -0 1829771 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@955 -- # uname 00:17:41.746 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:17:42.007 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1829771 00:17:42.007 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:17:42.007 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:17:42.007 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1829771' 00:17:42.007 killing process with pid 1829771 00:17:42.007 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@969 -- # kill 1829771 00:17:42.007 08:53:31 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@974 -- # wait 1829771 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@297 -- # iptr 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@789 -- # iptables-save 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@789 -- # iptables-restore 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@302 -- # remove_spdk_ns 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:17:42.007 08:53:32 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:17:44.555 00:17:44.555 real 0m12.960s 00:17:44.555 user 0m6.835s 00:17:44.555 sys 0m6.774s 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@1126 -- # xtrace_disable 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:17:44.555 ************************************ 00:17:44.555 END TEST nvmf_fused_ordering 00:17:44.555 ************************************ 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@26 -- # run_test nvmf_ns_masking test/nvmf/target/ns_masking.sh --transport=tcp 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:17:44.555 ************************************ 00:17:44.555 START TEST nvmf_ns_masking 00:17:44.555 ************************************ 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1125 -- # test/nvmf/target/ns_masking.sh --transport=tcp 00:17:44.555 * Looking for test storage... 00:17:44.555 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1689 -- # lcov --version 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@333 -- # local ver1 ver1_l 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@334 -- # local ver2 ver2_l 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@336 -- # IFS=.-: 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@336 -- # read -ra ver1 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@337 -- # IFS=.-: 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@337 -- # read -ra ver2 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@338 -- # local 'op=<' 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@340 -- # ver1_l=2 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@341 -- # ver2_l=1 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@344 -- # case "$op" in 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@345 -- # : 1 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@364 -- # (( v = 0 )) 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@365 -- # decimal 1 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@353 -- # local d=1 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@355 -- # echo 1 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@365 -- # ver1[v]=1 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@366 -- # decimal 2 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@353 -- # local d=2 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@355 -- # echo 2 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@366 -- # ver2[v]=2 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@368 -- # return 0 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:17:44.555 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:44.555 --rc genhtml_branch_coverage=1 00:17:44.555 --rc genhtml_function_coverage=1 00:17:44.555 --rc genhtml_legend=1 00:17:44.555 --rc geninfo_all_blocks=1 00:17:44.555 --rc geninfo_unexecuted_blocks=1 00:17:44.555 00:17:44.555 ' 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:17:44.555 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:44.555 --rc genhtml_branch_coverage=1 00:17:44.555 --rc genhtml_function_coverage=1 00:17:44.555 --rc genhtml_legend=1 00:17:44.555 --rc geninfo_all_blocks=1 00:17:44.555 --rc geninfo_unexecuted_blocks=1 00:17:44.555 00:17:44.555 ' 00:17:44.555 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:17:44.555 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:44.555 --rc genhtml_branch_coverage=1 00:17:44.555 --rc genhtml_function_coverage=1 00:17:44.555 --rc genhtml_legend=1 00:17:44.556 --rc geninfo_all_blocks=1 00:17:44.556 --rc geninfo_unexecuted_blocks=1 00:17:44.556 00:17:44.556 ' 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:17:44.556 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:17:44.556 --rc genhtml_branch_coverage=1 00:17:44.556 --rc genhtml_function_coverage=1 00:17:44.556 --rc genhtml_legend=1 00:17:44.556 --rc geninfo_all_blocks=1 00:17:44.556 --rc geninfo_unexecuted_blocks=1 00:17:44.556 00:17:44.556 ' 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@7 -- # uname -s 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@15 -- # shopt -s extglob 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@5 -- # export PATH 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@51 -- # : 0 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:17:44.556 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@55 -- # have_pci_nics=0 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@10 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@11 -- # hostsock=/var/tmp/host.sock 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@12 -- # loops=5 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@13 -- # uuidgen 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@13 -- # ns1uuid=55f89ffa-346a-4abe-a929-fe2ac5d02c4c 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@14 -- # uuidgen 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@14 -- # ns2uuid=88f7ec6f-d3c5-4462-97d0-1587ca36a4e1 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@16 -- # SUBSYSNQN=nqn.2016-06.io.spdk:cnode1 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@17 -- # HOSTNQN1=nqn.2016-06.io.spdk:host1 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@18 -- # HOSTNQN2=nqn.2016-06.io.spdk:host2 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@19 -- # uuidgen 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@19 -- # HOSTID=72bd57bd-66fc-4232-bcb6-a9e1e4e708a5 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@50 -- # nvmftestinit 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@474 -- # prepare_net_devs 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@436 -- # local -g is_hw=no 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@438 -- # remove_spdk_ns 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@309 -- # xtrace_disable 00:17:44.556 08:53:34 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@315 -- # pci_devs=() 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@315 -- # local -a pci_devs 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@316 -- # pci_net_devs=() 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@317 -- # pci_drivers=() 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@317 -- # local -A pci_drivers 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@319 -- # net_devs=() 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@319 -- # local -ga net_devs 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@320 -- # e810=() 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@320 -- # local -ga e810 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@321 -- # x722=() 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@321 -- # local -ga x722 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@322 -- # mlx=() 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@322 -- # local -ga mlx 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:17:52.701 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:17:52.701 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:17:52.702 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@416 -- # [[ up == up ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:17:52.702 Found net devices under 0000:4b:00.0: cvl_0_0 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@416 -- # [[ up == up ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:17:52.702 Found net devices under 0000:4b:00.1: cvl_0_1 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@440 -- # is_hw=yes 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:17:52.702 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:17:52.702 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.651 ms 00:17:52.702 00:17:52.702 --- 10.0.0.2 ping statistics --- 00:17:52.702 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:52.702 rtt min/avg/max/mdev = 0.651/0.651/0.651/0.000 ms 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:17:52.702 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:17:52.702 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.269 ms 00:17:52.702 00:17:52.702 --- 10.0.0.1 ping statistics --- 00:17:52.702 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:17:52.702 rtt min/avg/max/mdev = 0.269/0.269/0.269/0.000 ms 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@448 -- # return 0 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@51 -- # nvmfappstart 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@724 -- # xtrace_disable 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@507 -- # nvmfpid=1834488 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@508 -- # waitforlisten 1834488 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@831 -- # '[' -z 1834488 ']' 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@836 -- # local max_retries=100 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:52.702 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@840 -- # xtrace_disable 00:17:52.702 08:53:41 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:17:52.702 [2024-11-06 08:53:41.961641] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:17:52.702 [2024-11-06 08:53:41.961715] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:17:52.702 [2024-11-06 08:53:42.045005] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:17:52.702 [2024-11-06 08:53:42.086205] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:17:52.702 [2024-11-06 08:53:42.086243] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:17:52.702 [2024-11-06 08:53:42.086251] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:17:52.702 [2024-11-06 08:53:42.086257] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:17:52.703 [2024-11-06 08:53:42.086263] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:17:52.703 [2024-11-06 08:53:42.086898] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:17:52.703 08:53:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:17:52.703 08:53:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@864 -- # return 0 00:17:52.703 08:53:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:17:52.703 08:53:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@730 -- # xtrace_disable 00:17:52.703 08:53:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:17:52.703 08:53:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:17:52.703 08:53:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:17:52.964 [2024-11-06 08:53:42.942203] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:17:52.964 08:53:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@55 -- # MALLOC_BDEV_SIZE=64 00:17:52.964 08:53:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@56 -- # MALLOC_BLOCK_SIZE=512 00:17:52.964 08:53:42 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:17:53.224 Malloc1 00:17:53.224 08:53:43 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:17:53.485 Malloc2 00:17:53.485 08:53:43 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:17:53.485 08:53:43 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 00:17:53.746 08:53:43 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:17:54.007 [2024-11-06 08:53:43.873537] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:17:54.008 08:53:43 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@67 -- # connect 00:17:54.008 08:53:43 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 72bd57bd-66fc-4232-bcb6-a9e1e4e708a5 -a 10.0.0.2 -s 4420 -i 4 00:17:54.008 08:53:44 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 00:17:54.008 08:53:44 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1198 -- # local i=0 00:17:54.008 08:53:44 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:17:54.008 08:53:44 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:17:54.008 08:53:44 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # sleep 2 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # return 0 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@68 -- # ns_is_visible 0x1 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:17:56.553 [ 0]:0x1 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=34647fce43714fc9ab234c7c5092969c 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 34647fce43714fc9ab234c7c5092969c != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@72 -- # ns_is_visible 0x1 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:17:56.553 [ 0]:0x1 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=34647fce43714fc9ab234c7c5092969c 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 34647fce43714fc9ab234c7c5092969c != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@73 -- # ns_is_visible 0x2 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:17:56.553 [ 1]:0x2 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=38a1d748aefa4b65b41a2f1797822483 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 38a1d748aefa4b65b41a2f1797822483 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@75 -- # disconnect 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:17:56.553 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:17:56.553 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:17:56.813 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 --no-auto-visible 00:17:57.074 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@83 -- # connect 1 00:17:57.074 08:53:46 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 72bd57bd-66fc-4232-bcb6-a9e1e4e708a5 -a 10.0.0.2 -s 4420 -i 4 00:17:57.074 08:53:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 1 00:17:57.074 08:53:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1198 -- # local i=0 00:17:57.074 08:53:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:17:57.074 08:53:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1200 -- # [[ -n 1 ]] 00:17:57.074 08:53:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1201 -- # nvme_device_counter=1 00:17:57.074 08:53:47 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # sleep 2 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # return 0 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@84 -- # NOT ns_is_visible 0x1 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg ns_is_visible 0x1 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=ns_is_visible 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t ns_is_visible 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # ns_is_visible 0x1 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@85 -- # ns_is_visible 0x2 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:17:59.619 [ 0]:0x2 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=38a1d748aefa4b65b41a2f1797822483 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 38a1d748aefa4b65b41a2f1797822483 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@89 -- # ns_is_visible 0x1 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:17:59.619 [ 0]:0x1 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=34647fce43714fc9ab234c7c5092969c 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 34647fce43714fc9ab234c7c5092969c != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@90 -- # ns_is_visible 0x2 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:17:59.619 [ 1]:0x2 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=38a1d748aefa4b65b41a2f1797822483 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 38a1d748aefa4b65b41a2f1797822483 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:17:59.619 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@94 -- # NOT ns_is_visible 0x1 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg ns_is_visible 0x1 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=ns_is_visible 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t ns_is_visible 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # ns_is_visible 0x1 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@95 -- # ns_is_visible 0x2 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:17:59.880 [ 0]:0x2 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:17:59.880 08:53:49 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:18:00.140 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=38a1d748aefa4b65b41a2f1797822483 00:18:00.140 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 38a1d748aefa4b65b41a2f1797822483 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:18:00.140 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@97 -- # disconnect 00:18:00.140 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:18:00.140 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:00.140 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:18:00.140 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@101 -- # connect 2 00:18:00.140 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 72bd57bd-66fc-4232-bcb6-a9e1e4e708a5 -a 10.0.0.2 -s 4420 -i 4 00:18:00.401 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 2 00:18:00.401 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1198 -- # local i=0 00:18:00.401 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:18:00.401 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1200 -- # [[ -n 2 ]] 00:18:00.401 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1201 -- # nvme_device_counter=2 00:18:00.401 08:53:50 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # sleep 2 00:18:02.313 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:18:02.313 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:18:02.313 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:18:02.313 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # nvme_devices=2 00:18:02.313 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:18:02.313 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # return 0 00:18:02.313 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:18:02.313 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@102 -- # ns_is_visible 0x1 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:18:02.573 [ 0]:0x1 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=34647fce43714fc9ab234c7c5092969c 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 34647fce43714fc9ab234c7c5092969c != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@103 -- # ns_is_visible 0x2 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:18:02.573 [ 1]:0x2 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:18:02.573 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:18:02.574 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=38a1d748aefa4b65b41a2f1797822483 00:18:02.574 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 38a1d748aefa4b65b41a2f1797822483 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:18:02.574 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@106 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@107 -- # NOT ns_is_visible 0x1 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg ns_is_visible 0x1 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=ns_is_visible 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t ns_is_visible 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # ns_is_visible 0x1 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@108 -- # ns_is_visible 0x2 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:18:02.834 [ 0]:0x2 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=38a1d748aefa4b65b41a2f1797822483 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 38a1d748aefa4b65b41a2f1797822483 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@111 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:18:02.834 08:53:52 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:18:03.095 [2024-11-06 08:53:53.047964] nvmf_rpc.c:1870:nvmf_rpc_ns_visible_paused: *ERROR*: Unable to add/remove nqn.2016-06.io.spdk:host1 to namespace ID 2 00:18:03.095 request: 00:18:03.095 { 00:18:03.095 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:18:03.095 "nsid": 2, 00:18:03.095 "host": "nqn.2016-06.io.spdk:host1", 00:18:03.095 "method": "nvmf_ns_remove_host", 00:18:03.095 "req_id": 1 00:18:03.095 } 00:18:03.095 Got JSON-RPC error response 00:18:03.095 response: 00:18:03.095 { 00:18:03.095 "code": -32602, 00:18:03.095 "message": "Invalid parameters" 00:18:03.095 } 00:18:03.095 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:18:03.095 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:03.095 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:03.095 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:03.095 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@112 -- # NOT ns_is_visible 0x1 00:18:03.095 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg ns_is_visible 0x1 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=ns_is_visible 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t ns_is_visible 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # ns_is_visible 0x1 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@113 -- # ns_is_visible 0x2 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:18:03.096 [ 0]:0x2 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=38a1d748aefa4b65b41a2f1797822483 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 38a1d748aefa4b65b41a2f1797822483 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@114 -- # disconnect 00:18:03.096 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:18:03.356 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:03.356 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@118 -- # hostpid=1836982 00:18:03.356 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@117 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -r /var/tmp/host.sock -m 2 00:18:03.356 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@119 -- # trap 'killprocess $hostpid; nvmftestfini' SIGINT SIGTERM EXIT 00:18:03.356 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@121 -- # waitforlisten 1836982 /var/tmp/host.sock 00:18:03.356 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@831 -- # '[' -z 1836982 ']' 00:18:03.356 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/host.sock 00:18:03.356 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:03.356 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock...' 00:18:03.356 Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock... 00:18:03.356 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:03.356 08:53:53 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:18:03.356 [2024-11-06 08:53:53.305058] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:18:03.356 [2024-11-06 08:53:53.305112] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1836982 ] 00:18:03.356 [2024-11-06 08:53:53.392181] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:03.356 [2024-11-06 08:53:53.427850] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:04.297 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:04.297 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@864 -- # return 0 00:18:04.297 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@122 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:18:04.297 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:18:04.562 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@124 -- # uuid2nguid 55f89ffa-346a-4abe-a929-fe2ac5d02c4c 00:18:04.562 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@785 -- # tr -d - 00:18:04.562 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@124 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 -g 55F89FFA346A4ABEA929FE2AC5D02C4C -i 00:18:04.562 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@125 -- # uuid2nguid 88f7ec6f-d3c5-4462-97d0-1587ca36a4e1 00:18:04.562 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@785 -- # tr -d - 00:18:04.562 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 -g 88F7EC6FD3C5446297D01587CA36A4E1 -i 00:18:04.823 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@126 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:18:04.823 08:53:54 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@127 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host2 00:18:05.085 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@129 -- # hostrpc bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -b nvme0 00:18:05.085 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -b nvme0 00:18:05.347 nvme0n1 00:18:05.347 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@131 -- # hostrpc bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 -b nvme1 00:18:05.347 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 -b nvme1 00:18:05.919 nvme1n2 00:18:05.919 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # hostrpc bdev_get_bdevs 00:18:05.919 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs 00:18:05.919 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # jq -r '.[].name' 00:18:05.919 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # sort 00:18:05.919 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # xargs 00:18:05.919 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@134 -- # [[ nvme0n1 nvme1n2 == \n\v\m\e\0\n\1\ \n\v\m\e\1\n\2 ]] 00:18:05.919 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # hostrpc bdev_get_bdevs -b nvme0n1 00:18:05.919 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # jq -r '.[].uuid' 00:18:05.919 08:53:55 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs -b nvme0n1 00:18:06.179 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@135 -- # [[ 55f89ffa-346a-4abe-a929-fe2ac5d02c4c == \5\5\f\8\9\f\f\a\-\3\4\6\a\-\4\a\b\e\-\a\9\2\9\-\f\e\2\a\c\5\d\0\2\c\4\c ]] 00:18:06.179 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # hostrpc bdev_get_bdevs -b nvme1n2 00:18:06.179 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # jq -r '.[].uuid' 00:18:06.179 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs -b nvme1n2 00:18:06.179 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@136 -- # [[ 88f7ec6f-d3c5-4462-97d0-1587ca36a4e1 == \8\8\f\7\e\c\6\f\-\d\3\c\5\-\4\4\6\2\-\9\7\d\0\-\1\5\8\7\c\a\3\6\a\4\e\1 ]] 00:18:06.179 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@137 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:18:06.440 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@138 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:18:06.701 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@141 -- # uuid2nguid 55f89ffa-346a-4abe-a929-fe2ac5d02c4c 00:18:06.701 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@785 -- # tr -d - 00:18:06.701 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@141 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 invalid -n 1 -g 55F89FFA346A4ABEA929FE2AC5D02C4C 00:18:06.701 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@650 -- # local es=0 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 invalid -n 1 -g 55F89FFA346A4ABEA929FE2AC5D02C4C 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 invalid -n 1 -g 55F89FFA346A4ABEA929FE2AC5D02C4C 00:18:06.702 [2024-11-06 08:53:56.786272] bdev.c:8607:bdev_open_ext: *NOTICE*: Currently unable to find bdev with name: invalid 00:18:06.702 [2024-11-06 08:53:56.786305] subsystem.c:2151:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode1: bdev invalid cannot be opened, error=-19 00:18:06.702 [2024-11-06 08:53:56.786314] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:18:06.702 request: 00:18:06.702 { 00:18:06.702 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:18:06.702 "namespace": { 00:18:06.702 "bdev_name": "invalid", 00:18:06.702 "nsid": 1, 00:18:06.702 "nguid": "55F89FFA346A4ABEA929FE2AC5D02C4C", 00:18:06.702 "no_auto_visible": false 00:18:06.702 }, 00:18:06.702 "method": "nvmf_subsystem_add_ns", 00:18:06.702 "req_id": 1 00:18:06.702 } 00:18:06.702 Got JSON-RPC error response 00:18:06.702 response: 00:18:06.702 { 00:18:06.702 "code": -32602, 00:18:06.702 "message": "Invalid parameters" 00:18:06.702 } 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@653 -- # es=1 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@142 -- # uuid2nguid 55f89ffa-346a-4abe-a929-fe2ac5d02c4c 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@785 -- # tr -d - 00:18:06.702 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@142 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 -g 55F89FFA346A4ABEA929FE2AC5D02C4C -i 00:18:06.962 08:53:56 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@143 -- # sleep 2s 00:18:08.875 08:53:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@144 -- # jq length 00:18:08.875 08:53:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@144 -- # hostrpc bdev_get_bdevs 00:18:08.875 08:53:58 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs 00:18:09.135 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@144 -- # (( 0 == 0 )) 00:18:09.135 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@146 -- # killprocess 1836982 00:18:09.135 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@950 -- # '[' -z 1836982 ']' 00:18:09.135 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@954 -- # kill -0 1836982 00:18:09.135 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@955 -- # uname 00:18:09.135 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:09.135 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1836982 00:18:09.135 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:18:09.136 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:18:09.136 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1836982' 00:18:09.136 killing process with pid 1836982 00:18:09.136 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@969 -- # kill 1836982 00:18:09.136 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@974 -- # wait 1836982 00:18:09.396 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@147 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@149 -- # trap - SIGINT SIGTERM EXIT 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- target/ns_masking.sh@150 -- # nvmftestfini 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@514 -- # nvmfcleanup 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@121 -- # sync 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@124 -- # set +e 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@125 -- # for i in {1..20} 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:18:09.657 rmmod nvme_tcp 00:18:09.657 rmmod nvme_fabrics 00:18:09.657 rmmod nvme_keyring 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@128 -- # set -e 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@129 -- # return 0 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@515 -- # '[' -n 1834488 ']' 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@516 -- # killprocess 1834488 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@950 -- # '[' -z 1834488 ']' 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@954 -- # kill -0 1834488 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@955 -- # uname 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1834488 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1834488' 00:18:09.657 killing process with pid 1834488 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@969 -- # kill 1834488 00:18:09.657 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@974 -- # wait 1834488 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@297 -- # iptr 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@789 -- # iptables-save 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@789 -- # iptables-restore 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@302 -- # remove_spdk_ns 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:18:09.919 08:53:59 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:12.466 08:54:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:18:12.466 00:18:12.466 real 0m27.750s 00:18:12.466 user 0m31.468s 00:18:12.466 sys 0m8.029s 00:18:12.466 08:54:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:12.466 08:54:01 nvmf_tcp.nvmf_target_extra.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:18:12.466 ************************************ 00:18:12.466 END TEST nvmf_ns_masking 00:18:12.466 ************************************ 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@27 -- # [[ 1 -eq 1 ]] 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@28 -- # run_test nvmf_nvme_cli /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:18:12.466 ************************************ 00:18:12.466 START TEST nvmf_nvme_cli 00:18:12.466 ************************************ 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:18:12.466 * Looking for test storage... 00:18:12.466 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1689 -- # lcov --version 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@333 -- # local ver1 ver1_l 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@334 -- # local ver2 ver2_l 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@336 -- # IFS=.-: 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@336 -- # read -ra ver1 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@337 -- # IFS=.-: 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@337 -- # read -ra ver2 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@338 -- # local 'op=<' 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@340 -- # ver1_l=2 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@341 -- # ver2_l=1 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@344 -- # case "$op" in 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@345 -- # : 1 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@364 -- # (( v = 0 )) 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@365 -- # decimal 1 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@353 -- # local d=1 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@355 -- # echo 1 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@365 -- # ver1[v]=1 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@366 -- # decimal 2 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@353 -- # local d=2 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@355 -- # echo 2 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@366 -- # ver2[v]=2 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@368 -- # return 0 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:12.466 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:18:12.466 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:12.466 --rc genhtml_branch_coverage=1 00:18:12.466 --rc genhtml_function_coverage=1 00:18:12.466 --rc genhtml_legend=1 00:18:12.466 --rc geninfo_all_blocks=1 00:18:12.466 --rc geninfo_unexecuted_blocks=1 00:18:12.466 00:18:12.467 ' 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:18:12.467 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:12.467 --rc genhtml_branch_coverage=1 00:18:12.467 --rc genhtml_function_coverage=1 00:18:12.467 --rc genhtml_legend=1 00:18:12.467 --rc geninfo_all_blocks=1 00:18:12.467 --rc geninfo_unexecuted_blocks=1 00:18:12.467 00:18:12.467 ' 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:18:12.467 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:12.467 --rc genhtml_branch_coverage=1 00:18:12.467 --rc genhtml_function_coverage=1 00:18:12.467 --rc genhtml_legend=1 00:18:12.467 --rc geninfo_all_blocks=1 00:18:12.467 --rc geninfo_unexecuted_blocks=1 00:18:12.467 00:18:12.467 ' 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:18:12.467 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:12.467 --rc genhtml_branch_coverage=1 00:18:12.467 --rc genhtml_function_coverage=1 00:18:12.467 --rc genhtml_legend=1 00:18:12.467 --rc geninfo_all_blocks=1 00:18:12.467 --rc geninfo_unexecuted_blocks=1 00:18:12.467 00:18:12.467 ' 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@7 -- # uname -s 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@15 -- # shopt -s extglob 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@5 -- # export PATH 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@51 -- # : 0 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:18:12.467 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@55 -- # have_pci_nics=0 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@11 -- # MALLOC_BDEV_SIZE=64 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@14 -- # devs=() 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@16 -- # nvmftestinit 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@474 -- # prepare_net_devs 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@436 -- # local -g is_hw=no 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@438 -- # remove_spdk_ns 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@309 -- # xtrace_disable 00:18:12.467 08:54:02 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@315 -- # pci_devs=() 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@315 -- # local -a pci_devs 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@316 -- # pci_net_devs=() 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@317 -- # pci_drivers=() 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@317 -- # local -A pci_drivers 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@319 -- # net_devs=() 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@319 -- # local -ga net_devs 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@320 -- # e810=() 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@320 -- # local -ga e810 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@321 -- # x722=() 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@321 -- # local -ga x722 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@322 -- # mlx=() 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@322 -- # local -ga mlx 00:18:20.617 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:18:20.618 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:18:20.618 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@416 -- # [[ up == up ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:18:20.618 Found net devices under 0000:4b:00.0: cvl_0_0 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@416 -- # [[ up == up ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:18:20.618 Found net devices under 0000:4b:00.1: cvl_0_1 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@440 -- # is_hw=yes 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:18:20.618 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:18:20.618 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.669 ms 00:18:20.618 00:18:20.618 --- 10.0.0.2 ping statistics --- 00:18:20.618 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:20.618 rtt min/avg/max/mdev = 0.669/0.669/0.669/0.000 ms 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:18:20.618 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:18:20.618 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.314 ms 00:18:20.618 00:18:20.618 --- 10.0.0.1 ping statistics --- 00:18:20.618 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:20.618 rtt min/avg/max/mdev = 0.314/0.314/0.314/0.000 ms 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@448 -- # return 0 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@17 -- # nvmfappstart -m 0xF 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@724 -- # xtrace_disable 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@507 -- # nvmfpid=1842734 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@508 -- # waitforlisten 1842734 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:18:20.618 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@831 -- # '[' -z 1842734 ']' 00:18:20.619 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:20.619 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:20.619 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:20.619 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:20.619 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:20.619 08:54:09 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.619 [2024-11-06 08:54:09.703643] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:18:20.619 [2024-11-06 08:54:09.703697] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:18:20.619 [2024-11-06 08:54:09.781894] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:18:20.619 [2024-11-06 08:54:09.820224] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:18:20.619 [2024-11-06 08:54:09.820257] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:18:20.619 [2024-11-06 08:54:09.820265] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:18:20.619 [2024-11-06 08:54:09.820272] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:18:20.619 [2024-11-06 08:54:09.820277] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:18:20.619 [2024-11-06 08:54:09.821804] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:20.619 [2024-11-06 08:54:09.821855] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:18:20.619 [2024-11-06 08:54:09.822017] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:20.619 [2024-11-06 08:54:09.822018] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@864 -- # return 0 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@730 -- # xtrace_disable 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@19 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.619 [2024-11-06 08:54:10.544900] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@21 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.619 Malloc0 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.619 Malloc1 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME -d SPDK_Controller1 -i 291 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.619 [2024-11-06 08:54:10.642407] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@28 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:20.619 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@30 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -a 10.0.0.2 -s 4420 00:18:20.880 00:18:20.880 Discovery Log Number of Records 2, Generation counter 2 00:18:20.880 =====Discovery Log Entry 0====== 00:18:20.880 trtype: tcp 00:18:20.880 adrfam: ipv4 00:18:20.880 subtype: current discovery subsystem 00:18:20.880 treq: not required 00:18:20.880 portid: 0 00:18:20.880 trsvcid: 4420 00:18:20.880 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:18:20.880 traddr: 10.0.0.2 00:18:20.880 eflags: explicit discovery connections, duplicate discovery information 00:18:20.880 sectype: none 00:18:20.880 =====Discovery Log Entry 1====== 00:18:20.880 trtype: tcp 00:18:20.880 adrfam: ipv4 00:18:20.880 subtype: nvme subsystem 00:18:20.880 treq: not required 00:18:20.880 portid: 0 00:18:20.880 trsvcid: 4420 00:18:20.880 subnqn: nqn.2016-06.io.spdk:cnode1 00:18:20.880 traddr: 10.0.0.2 00:18:20.880 eflags: none 00:18:20.880 sectype: none 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # devs=($(get_nvme_devs)) 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # get_nvme_devs 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@548 -- # local dev _ 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@547 -- # nvme list 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ Node == /dev/nvme* ]] 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ --------------------- == /dev/nvme* ]] 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # nvme_num_before_connection=0 00:18:20.880 08:54:10 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@32 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:18:22.266 08:54:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@34 -- # waitforserial SPDKISFASTANDAWESOME 2 00:18:22.266 08:54:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1198 -- # local i=0 00:18:22.266 08:54:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:18:22.266 08:54:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1200 -- # [[ -n 2 ]] 00:18:22.266 08:54:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1201 -- # nvme_device_counter=2 00:18:22.266 08:54:12 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1205 -- # sleep 2 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1207 -- # nvme_devices=2 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1208 -- # return 0 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@35 -- # get_nvme_devs 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@548 -- # local dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@547 -- # nvme list 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ Node == /dev/nvme* ]] 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ --------------------- == /dev/nvme* ]] 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # echo /dev/nvme0n1 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # echo /dev/nvme0n2 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@35 -- # [[ -z /dev/nvme0n1 00:18:24.847 /dev/nvme0n2 ]] 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # devs=($(get_nvme_devs)) 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # get_nvme_devs 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@548 -- # local dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@547 -- # nvme list 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ Node == /dev/nvme* ]] 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ --------------------- == /dev/nvme* ]] 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # echo /dev/nvme0n1 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@551 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@552 -- # echo /dev/nvme0n2 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@550 -- # read -r dev _ 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # nvme_num=2 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@60 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:18:24.847 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@61 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1219 -- # local i=0 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1231 -- # return 0 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@62 -- # (( nvme_num <= nvme_num_before_connection )) 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@67 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@561 -- # xtrace_disable 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- target/nvme_cli.sh@70 -- # nvmftestfini 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@514 -- # nvmfcleanup 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@121 -- # sync 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@124 -- # set +e 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@125 -- # for i in {1..20} 00:18:24.847 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:18:24.847 rmmod nvme_tcp 00:18:24.847 rmmod nvme_fabrics 00:18:24.847 rmmod nvme_keyring 00:18:25.109 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:18:25.109 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@128 -- # set -e 00:18:25.109 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@129 -- # return 0 00:18:25.109 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@515 -- # '[' -n 1842734 ']' 00:18:25.109 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@516 -- # killprocess 1842734 00:18:25.109 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@950 -- # '[' -z 1842734 ']' 00:18:25.109 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@954 -- # kill -0 1842734 00:18:25.109 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@955 -- # uname 00:18:25.109 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:18:25.109 08:54:14 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1842734 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1842734' 00:18:25.109 killing process with pid 1842734 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@969 -- # kill 1842734 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@974 -- # wait 1842734 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@297 -- # iptr 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@789 -- # iptables-save 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@789 -- # iptables-restore 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@302 -- # remove_spdk_ns 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:18:25.109 08:54:15 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:18:27.658 00:18:27.658 real 0m15.222s 00:18:27.658 user 0m23.558s 00:18:27.658 sys 0m6.253s 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@1126 -- # xtrace_disable 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:18:27.658 ************************************ 00:18:27.658 END TEST nvmf_nvme_cli 00:18:27.658 ************************************ 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@30 -- # [[ 1 -eq 1 ]] 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@31 -- # run_test nvmf_vfio_user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:18:27.658 ************************************ 00:18:27.658 START TEST nvmf_vfio_user 00:18:27.658 ************************************ 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:18:27.658 * Looking for test storage... 00:18:27.658 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1689 -- # lcov --version 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@333 -- # local ver1 ver1_l 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@334 -- # local ver2 ver2_l 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@336 -- # IFS=.-: 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@336 -- # read -ra ver1 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@337 -- # IFS=.-: 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@337 -- # read -ra ver2 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@338 -- # local 'op=<' 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@340 -- # ver1_l=2 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@341 -- # ver2_l=1 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@344 -- # case "$op" in 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@345 -- # : 1 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@364 -- # (( v = 0 )) 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@365 -- # decimal 1 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@353 -- # local d=1 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@355 -- # echo 1 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@365 -- # ver1[v]=1 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@366 -- # decimal 2 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@353 -- # local d=2 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@355 -- # echo 2 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@366 -- # ver2[v]=2 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@368 -- # return 0 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:18:27.658 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:27.658 --rc genhtml_branch_coverage=1 00:18:27.658 --rc genhtml_function_coverage=1 00:18:27.658 --rc genhtml_legend=1 00:18:27.658 --rc geninfo_all_blocks=1 00:18:27.658 --rc geninfo_unexecuted_blocks=1 00:18:27.658 00:18:27.658 ' 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:18:27.658 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:27.658 --rc genhtml_branch_coverage=1 00:18:27.658 --rc genhtml_function_coverage=1 00:18:27.658 --rc genhtml_legend=1 00:18:27.658 --rc geninfo_all_blocks=1 00:18:27.658 --rc geninfo_unexecuted_blocks=1 00:18:27.658 00:18:27.658 ' 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:18:27.658 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:27.658 --rc genhtml_branch_coverage=1 00:18:27.658 --rc genhtml_function_coverage=1 00:18:27.658 --rc genhtml_legend=1 00:18:27.658 --rc geninfo_all_blocks=1 00:18:27.658 --rc geninfo_unexecuted_blocks=1 00:18:27.658 00:18:27.658 ' 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:18:27.658 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:18:27.658 --rc genhtml_branch_coverage=1 00:18:27.658 --rc genhtml_function_coverage=1 00:18:27.658 --rc genhtml_legend=1 00:18:27.658 --rc geninfo_all_blocks=1 00:18:27.658 --rc geninfo_unexecuted_blocks=1 00:18:27.658 00:18:27.658 ' 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@7 -- # uname -s 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@15 -- # shopt -s extglob 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:18:27.658 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@5 -- # export PATH 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@51 -- # : 0 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:18:27.659 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- nvmf/common.sh@55 -- # have_pci_nics=0 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@12 -- # MALLOC_BDEV_SIZE=64 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@14 -- # NUM_DEVICES=2 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@47 -- # rm -rf /var/run/vfio-user 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@103 -- # setup_nvmf_vfio_user '' '' 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args= 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@52 -- # local transport_args= 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=1844749 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 1844749' 00:18:27.659 Process pid: 1844749 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 1844749 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@831 -- # '[' -z 1844749 ']' 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@836 -- # local max_retries=100 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:27.659 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@840 -- # xtrace_disable 00:18:27.659 08:54:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:18:27.659 [2024-11-06 08:54:17.644929] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:18:27.659 [2024-11-06 08:54:17.645003] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:18:27.659 [2024-11-06 08:54:17.722898] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:18:27.659 [2024-11-06 08:54:17.764868] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:18:27.659 [2024-11-06 08:54:17.764901] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:18:27.659 [2024-11-06 08:54:17.764909] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:18:27.659 [2024-11-06 08:54:17.764916] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:18:27.659 [2024-11-06 08:54:17.764922] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:18:27.659 [2024-11-06 08:54:17.766519] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:18:27.659 [2024-11-06 08:54:17.766638] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:18:27.659 [2024-11-06 08:54:17.766797] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:18:27.659 [2024-11-06 08:54:17.766962] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:18:28.601 08:54:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:18:28.601 08:54:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@864 -- # return 0 00:18:28.601 08:54:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:18:29.545 08:54:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER 00:18:29.545 08:54:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:18:29.545 08:54:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:18:29.545 08:54:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:18:29.545 08:54:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:18:29.545 08:54:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:18:29.804 Malloc1 00:18:29.804 08:54:19 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:18:30.065 08:54:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:18:30.325 08:54:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:18:30.325 08:54:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:18:30.325 08:54:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:18:30.325 08:54:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:18:30.585 Malloc2 00:18:30.585 08:54:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:18:30.845 08:54:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:18:30.845 08:54:20 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:18:31.105 08:54:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@104 -- # run_nvmf_vfio_user 00:18:31.105 08:54:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # seq 1 2 00:18:31.105 08:54:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:18:31.105 08:54:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user1/1 00:18:31.105 08:54:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode1 00:18:31.106 08:54:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -L nvme -L nvme_vfio -L vfio_pci 00:18:31.106 [2024-11-06 08:54:21.159319] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:18:31.106 [2024-11-06 08:54:21.159389] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1845448 ] 00:18:31.106 [2024-11-06 08:54:21.211568] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user1/1 00:18:31.368 [2024-11-06 08:54:21.220034] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:18:31.368 [2024-11-06 08:54:21.220054] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7f7029ed6000 00:18:31.368 [2024-11-06 08:54:21.221030] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:18:31.368 [2024-11-06 08:54:21.222037] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:18:31.368 [2024-11-06 08:54:21.223036] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:18:31.368 [2024-11-06 08:54:21.224049] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:18:31.368 [2024-11-06 08:54:21.225051] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:18:31.368 [2024-11-06 08:54:21.226049] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:18:31.368 [2024-11-06 08:54:21.227057] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:18:31.368 [2024-11-06 08:54:21.228063] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:18:31.368 [2024-11-06 08:54:21.229075] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:18:31.368 [2024-11-06 08:54:21.229088] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7f7029ecb000 00:18:31.368 [2024-11-06 08:54:21.230416] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:18:31.368 [2024-11-06 08:54:21.251334] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user1/1/cntrl Setup Successfully 00:18:31.368 [2024-11-06 08:54:21.251372] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to connect adminq (no timeout) 00:18:31.368 [2024-11-06 08:54:21.254223] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:18:31.368 [2024-11-06 08:54:21.254268] nvme_pcie_common.c: 134:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:18:31.368 [2024-11-06 08:54:21.254353] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for connect adminq (no timeout) 00:18:31.368 [2024-11-06 08:54:21.254372] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read vs (no timeout) 00:18:31.368 [2024-11-06 08:54:21.254377] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read vs wait for vs (no timeout) 00:18:31.368 [2024-11-06 08:54:21.255223] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x8, value 0x10300 00:18:31.368 [2024-11-06 08:54:21.255232] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read cap (no timeout) 00:18:31.368 [2024-11-06 08:54:21.255240] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to read cap wait for cap (no timeout) 00:18:31.368 [2024-11-06 08:54:21.256221] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:18:31.368 [2024-11-06 08:54:21.256230] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to check en (no timeout) 00:18:31.368 [2024-11-06 08:54:21.256238] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to check en wait for cc (timeout 15000 ms) 00:18:31.368 [2024-11-06 08:54:21.257227] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x0 00:18:31.368 [2024-11-06 08:54:21.257236] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:18:31.368 [2024-11-06 08:54:21.258237] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x0 00:18:31.368 [2024-11-06 08:54:21.258246] nvme_ctrlr.c:3870:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] CC.EN = 0 && CSTS.RDY = 0 00:18:31.368 [2024-11-06 08:54:21.258251] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to controller is disabled (timeout 15000 ms) 00:18:31.368 [2024-11-06 08:54:21.258260] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:18:31.368 [2024-11-06 08:54:21.258366] nvme_ctrlr.c:4068:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Setting CC.EN = 1 00:18:31.368 [2024-11-06 08:54:21.258371] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:18:31.368 [2024-11-06 08:54:21.258376] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x28, value 0x2000003c0000 00:18:31.368 [2024-11-06 08:54:21.259248] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x30, value 0x2000003be000 00:18:31.369 [2024-11-06 08:54:21.260251] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x24, value 0xff00ff 00:18:31.369 [2024-11-06 08:54:21.261258] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:18:31.369 [2024-11-06 08:54:21.262251] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:18:31.369 [2024-11-06 08:54:21.262305] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:18:31.369 [2024-11-06 08:54:21.263266] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x1 00:18:31.369 [2024-11-06 08:54:21.263274] nvme_ctrlr.c:3905:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:18:31.369 [2024-11-06 08:54:21.263279] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to reset admin queue (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263300] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify controller (no timeout) 00:18:31.369 [2024-11-06 08:54:21.263314] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify controller (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263330] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:18:31.369 [2024-11-06 08:54:21.263336] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:18:31.369 [2024-11-06 08:54:21.263340] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:31.369 [2024-11-06 08:54:21.263354] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263392] nvme_ctrlr.c:2054:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] transport max_xfer_size 131072 00:18:31.369 [2024-11-06 08:54:21.263397] nvme_ctrlr.c:2058:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] MDTS max_xfer_size 131072 00:18:31.369 [2024-11-06 08:54:21.263402] nvme_ctrlr.c:2061:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] CNTLID 0x0001 00:18:31.369 [2024-11-06 08:54:21.263407] nvme_ctrlr.c:2072:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:18:31.369 [2024-11-06 08:54:21.263413] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] transport max_sges 1 00:18:31.369 [2024-11-06 08:54:21.263417] nvme_ctrlr.c:2100:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] fuses compare and write: 1 00:18:31.369 [2024-11-06 08:54:21.263424] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to configure AER (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263433] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for configure aer (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263443] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263471] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:18:31.369 [2024-11-06 08:54:21.263480] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:18:31.369 [2024-11-06 08:54:21.263488] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:18:31.369 [2024-11-06 08:54:21.263497] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:18:31.369 [2024-11-06 08:54:21.263502] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set keep alive timeout (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263509] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263518] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263533] nvme_ctrlr.c:3011:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Controller adjusted keep alive timeout to 0 ms 00:18:31.369 [2024-11-06 08:54:21.263538] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify controller iocs specific (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263546] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set number of queues (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263552] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for set number of queues (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263561] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263636] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify active ns (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263645] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify active ns (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263652] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:18:31.369 [2024-11-06 08:54:21.263657] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:18:31.369 [2024-11-06 08:54:21.263660] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:31.369 [2024-11-06 08:54:21.263667] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263692] nvme_ctrlr.c:4699:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Namespace 1 was added 00:18:31.369 [2024-11-06 08:54:21.263700] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify ns (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263708] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify ns (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263716] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:18:31.369 [2024-11-06 08:54:21.263720] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:18:31.369 [2024-11-06 08:54:21.263723] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:31.369 [2024-11-06 08:54:21.263729] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263763] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify namespace id descriptors (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263772] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263779] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:18:31.369 [2024-11-06 08:54:21.263783] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:18:31.369 [2024-11-06 08:54:21.263787] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:31.369 [2024-11-06 08:54:21.263793] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263813] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to identify ns iocs specific (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263820] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set supported log pages (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263828] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set supported features (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263834] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set host behavior support feature (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263839] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set doorbell buffer config (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263844] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to set host ID (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263850] nvme_ctrlr.c:3111:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] NVMe-oF transport - not sending Set Features - Host ID 00:18:31.369 [2024-11-06 08:54:21.263854] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to transport ready (timeout 30000 ms) 00:18:31.369 [2024-11-06 08:54:21.263859] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] setting state to ready (no timeout) 00:18:31.369 [2024-11-06 08:54:21.263878] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263902] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263921] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263942] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:18:31.369 [2024-11-06 08:54:21.263952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:18:31.369 [2024-11-06 08:54:21.263965] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:18:31.369 [2024-11-06 08:54:21.263969] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:18:31.369 [2024-11-06 08:54:21.263973] nvme_pcie_common.c:1241:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:18:31.370 [2024-11-06 08:54:21.263977] nvme_pcie_common.c:1257:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:18:31.370 [2024-11-06 08:54:21.263980] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 2 00:18:31.370 [2024-11-06 08:54:21.263986] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:18:31.370 [2024-11-06 08:54:21.263994] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:18:31.370 [2024-11-06 08:54:21.263998] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:18:31.370 [2024-11-06 08:54:21.264002] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:31.370 [2024-11-06 08:54:21.264008] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:18:31.370 [2024-11-06 08:54:21.264015] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:18:31.370 [2024-11-06 08:54:21.264020] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:18:31.370 [2024-11-06 08:54:21.264023] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:31.370 [2024-11-06 08:54:21.264029] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:18:31.370 [2024-11-06 08:54:21.264039] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:18:31.370 [2024-11-06 08:54:21.264043] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:18:31.370 [2024-11-06 08:54:21.264047] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:31.370 [2024-11-06 08:54:21.264053] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:18:31.370 [2024-11-06 08:54:21.264060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:18:31.370 [2024-11-06 08:54:21.264071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:18:31.370 [2024-11-06 08:54:21.264082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:18:31.370 [2024-11-06 08:54:21.264089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:18:31.370 ===================================================== 00:18:31.370 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:18:31.370 ===================================================== 00:18:31.370 Controller Capabilities/Features 00:18:31.370 ================================ 00:18:31.370 Vendor ID: 4e58 00:18:31.370 Subsystem Vendor ID: 4e58 00:18:31.370 Serial Number: SPDK1 00:18:31.370 Model Number: SPDK bdev Controller 00:18:31.370 Firmware Version: 25.01 00:18:31.370 Recommended Arb Burst: 6 00:18:31.370 IEEE OUI Identifier: 8d 6b 50 00:18:31.370 Multi-path I/O 00:18:31.370 May have multiple subsystem ports: Yes 00:18:31.370 May have multiple controllers: Yes 00:18:31.370 Associated with SR-IOV VF: No 00:18:31.370 Max Data Transfer Size: 131072 00:18:31.370 Max Number of Namespaces: 32 00:18:31.370 Max Number of I/O Queues: 127 00:18:31.370 NVMe Specification Version (VS): 1.3 00:18:31.370 NVMe Specification Version (Identify): 1.3 00:18:31.370 Maximum Queue Entries: 256 00:18:31.370 Contiguous Queues Required: Yes 00:18:31.370 Arbitration Mechanisms Supported 00:18:31.370 Weighted Round Robin: Not Supported 00:18:31.370 Vendor Specific: Not Supported 00:18:31.370 Reset Timeout: 15000 ms 00:18:31.370 Doorbell Stride: 4 bytes 00:18:31.370 NVM Subsystem Reset: Not Supported 00:18:31.370 Command Sets Supported 00:18:31.370 NVM Command Set: Supported 00:18:31.370 Boot Partition: Not Supported 00:18:31.370 Memory Page Size Minimum: 4096 bytes 00:18:31.370 Memory Page Size Maximum: 4096 bytes 00:18:31.370 Persistent Memory Region: Not Supported 00:18:31.370 Optional Asynchronous Events Supported 00:18:31.370 Namespace Attribute Notices: Supported 00:18:31.370 Firmware Activation Notices: Not Supported 00:18:31.370 ANA Change Notices: Not Supported 00:18:31.370 PLE Aggregate Log Change Notices: Not Supported 00:18:31.370 LBA Status Info Alert Notices: Not Supported 00:18:31.370 EGE Aggregate Log Change Notices: Not Supported 00:18:31.370 Normal NVM Subsystem Shutdown event: Not Supported 00:18:31.370 Zone Descriptor Change Notices: Not Supported 00:18:31.370 Discovery Log Change Notices: Not Supported 00:18:31.370 Controller Attributes 00:18:31.370 128-bit Host Identifier: Supported 00:18:31.370 Non-Operational Permissive Mode: Not Supported 00:18:31.370 NVM Sets: Not Supported 00:18:31.370 Read Recovery Levels: Not Supported 00:18:31.370 Endurance Groups: Not Supported 00:18:31.370 Predictable Latency Mode: Not Supported 00:18:31.370 Traffic Based Keep ALive: Not Supported 00:18:31.370 Namespace Granularity: Not Supported 00:18:31.370 SQ Associations: Not Supported 00:18:31.370 UUID List: Not Supported 00:18:31.370 Multi-Domain Subsystem: Not Supported 00:18:31.370 Fixed Capacity Management: Not Supported 00:18:31.370 Variable Capacity Management: Not Supported 00:18:31.370 Delete Endurance Group: Not Supported 00:18:31.370 Delete NVM Set: Not Supported 00:18:31.370 Extended LBA Formats Supported: Not Supported 00:18:31.370 Flexible Data Placement Supported: Not Supported 00:18:31.370 00:18:31.370 Controller Memory Buffer Support 00:18:31.370 ================================ 00:18:31.370 Supported: No 00:18:31.370 00:18:31.370 Persistent Memory Region Support 00:18:31.370 ================================ 00:18:31.370 Supported: No 00:18:31.370 00:18:31.370 Admin Command Set Attributes 00:18:31.370 ============================ 00:18:31.370 Security Send/Receive: Not Supported 00:18:31.370 Format NVM: Not Supported 00:18:31.370 Firmware Activate/Download: Not Supported 00:18:31.370 Namespace Management: Not Supported 00:18:31.370 Device Self-Test: Not Supported 00:18:31.370 Directives: Not Supported 00:18:31.370 NVMe-MI: Not Supported 00:18:31.370 Virtualization Management: Not Supported 00:18:31.370 Doorbell Buffer Config: Not Supported 00:18:31.370 Get LBA Status Capability: Not Supported 00:18:31.370 Command & Feature Lockdown Capability: Not Supported 00:18:31.370 Abort Command Limit: 4 00:18:31.370 Async Event Request Limit: 4 00:18:31.370 Number of Firmware Slots: N/A 00:18:31.370 Firmware Slot 1 Read-Only: N/A 00:18:31.370 Firmware Activation Without Reset: N/A 00:18:31.370 Multiple Update Detection Support: N/A 00:18:31.370 Firmware Update Granularity: No Information Provided 00:18:31.370 Per-Namespace SMART Log: No 00:18:31.370 Asymmetric Namespace Access Log Page: Not Supported 00:18:31.370 Subsystem NQN: nqn.2019-07.io.spdk:cnode1 00:18:31.370 Command Effects Log Page: Supported 00:18:31.370 Get Log Page Extended Data: Supported 00:18:31.370 Telemetry Log Pages: Not Supported 00:18:31.370 Persistent Event Log Pages: Not Supported 00:18:31.370 Supported Log Pages Log Page: May Support 00:18:31.370 Commands Supported & Effects Log Page: Not Supported 00:18:31.370 Feature Identifiers & Effects Log Page:May Support 00:18:31.370 NVMe-MI Commands & Effects Log Page: May Support 00:18:31.370 Data Area 4 for Telemetry Log: Not Supported 00:18:31.370 Error Log Page Entries Supported: 128 00:18:31.370 Keep Alive: Supported 00:18:31.370 Keep Alive Granularity: 10000 ms 00:18:31.370 00:18:31.370 NVM Command Set Attributes 00:18:31.370 ========================== 00:18:31.370 Submission Queue Entry Size 00:18:31.370 Max: 64 00:18:31.370 Min: 64 00:18:31.370 Completion Queue Entry Size 00:18:31.370 Max: 16 00:18:31.370 Min: 16 00:18:31.370 Number of Namespaces: 32 00:18:31.370 Compare Command: Supported 00:18:31.370 Write Uncorrectable Command: Not Supported 00:18:31.370 Dataset Management Command: Supported 00:18:31.370 Write Zeroes Command: Supported 00:18:31.370 Set Features Save Field: Not Supported 00:18:31.370 Reservations: Not Supported 00:18:31.370 Timestamp: Not Supported 00:18:31.370 Copy: Supported 00:18:31.370 Volatile Write Cache: Present 00:18:31.370 Atomic Write Unit (Normal): 1 00:18:31.370 Atomic Write Unit (PFail): 1 00:18:31.370 Atomic Compare & Write Unit: 1 00:18:31.370 Fused Compare & Write: Supported 00:18:31.370 Scatter-Gather List 00:18:31.370 SGL Command Set: Supported (Dword aligned) 00:18:31.370 SGL Keyed: Not Supported 00:18:31.370 SGL Bit Bucket Descriptor: Not Supported 00:18:31.370 SGL Metadata Pointer: Not Supported 00:18:31.370 Oversized SGL: Not Supported 00:18:31.370 SGL Metadata Address: Not Supported 00:18:31.370 SGL Offset: Not Supported 00:18:31.370 Transport SGL Data Block: Not Supported 00:18:31.370 Replay Protected Memory Block: Not Supported 00:18:31.370 00:18:31.370 Firmware Slot Information 00:18:31.370 ========================= 00:18:31.370 Active slot: 1 00:18:31.370 Slot 1 Firmware Revision: 25.01 00:18:31.370 00:18:31.370 00:18:31.370 Commands Supported and Effects 00:18:31.370 ============================== 00:18:31.370 Admin Commands 00:18:31.370 -------------- 00:18:31.370 Get Log Page (02h): Supported 00:18:31.370 Identify (06h): Supported 00:18:31.370 Abort (08h): Supported 00:18:31.370 Set Features (09h): Supported 00:18:31.370 Get Features (0Ah): Supported 00:18:31.370 Asynchronous Event Request (0Ch): Supported 00:18:31.370 Keep Alive (18h): Supported 00:18:31.370 I/O Commands 00:18:31.370 ------------ 00:18:31.371 Flush (00h): Supported LBA-Change 00:18:31.371 Write (01h): Supported LBA-Change 00:18:31.371 Read (02h): Supported 00:18:31.371 Compare (05h): Supported 00:18:31.371 Write Zeroes (08h): Supported LBA-Change 00:18:31.371 Dataset Management (09h): Supported LBA-Change 00:18:31.371 Copy (19h): Supported LBA-Change 00:18:31.371 00:18:31.371 Error Log 00:18:31.371 ========= 00:18:31.371 00:18:31.371 Arbitration 00:18:31.371 =========== 00:18:31.371 Arbitration Burst: 1 00:18:31.371 00:18:31.371 Power Management 00:18:31.371 ================ 00:18:31.371 Number of Power States: 1 00:18:31.371 Current Power State: Power State #0 00:18:31.371 Power State #0: 00:18:31.371 Max Power: 0.00 W 00:18:31.371 Non-Operational State: Operational 00:18:31.371 Entry Latency: Not Reported 00:18:31.371 Exit Latency: Not Reported 00:18:31.371 Relative Read Throughput: 0 00:18:31.371 Relative Read Latency: 0 00:18:31.371 Relative Write Throughput: 0 00:18:31.371 Relative Write Latency: 0 00:18:31.371 Idle Power: Not Reported 00:18:31.371 Active Power: Not Reported 00:18:31.371 Non-Operational Permissive Mode: Not Supported 00:18:31.371 00:18:31.371 Health Information 00:18:31.371 ================== 00:18:31.371 Critical Warnings: 00:18:31.371 Available Spare Space: OK 00:18:31.371 Temperature: OK 00:18:31.371 Device Reliability: OK 00:18:31.371 Read Only: No 00:18:31.371 Volatile Memory Backup: OK 00:18:31.371 Current Temperature: 0 Kelvin (-273 Celsius) 00:18:31.371 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:18:31.371 Available Spare: 0% 00:18:31.371 Available Sp[2024-11-06 08:54:21.264190] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:18:31.371 [2024-11-06 08:54:21.264199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:18:31.371 [2024-11-06 08:54:21.264228] nvme_ctrlr.c:4363:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] Prepare to destruct SSD 00:18:31.371 [2024-11-06 08:54:21.264238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:18:31.371 [2024-11-06 08:54:21.264244] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:18:31.371 [2024-11-06 08:54:21.264251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:18:31.371 [2024-11-06 08:54:21.264257] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:18:31.371 [2024-11-06 08:54:21.265268] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:18:31.371 [2024-11-06 08:54:21.265279] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x464001 00:18:31.371 [2024-11-06 08:54:21.266268] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:18:31.371 [2024-11-06 08:54:21.266309] nvme_ctrlr.c:1124:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] RTD3E = 0 us 00:18:31.371 [2024-11-06 08:54:21.266315] nvme_ctrlr.c:1127:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] shutdown timeout = 10000 ms 00:18:31.371 [2024-11-06 08:54:21.267276] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x9 00:18:31.371 [2024-11-06 08:54:21.267288] nvme_ctrlr.c:1246:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1, 0] shutdown complete in 0 milliseconds 00:18:31.371 [2024-11-06 08:54:21.267349] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user1/1/cntrl 00:18:31.371 [2024-11-06 08:54:21.270755] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:18:31.371 are Threshold: 0% 00:18:31.371 Life Percentage Used: 0% 00:18:31.371 Data Units Read: 0 00:18:31.371 Data Units Written: 0 00:18:31.371 Host Read Commands: 0 00:18:31.371 Host Write Commands: 0 00:18:31.371 Controller Busy Time: 0 minutes 00:18:31.371 Power Cycles: 0 00:18:31.371 Power On Hours: 0 hours 00:18:31.371 Unsafe Shutdowns: 0 00:18:31.371 Unrecoverable Media Errors: 0 00:18:31.371 Lifetime Error Log Entries: 0 00:18:31.371 Warning Temperature Time: 0 minutes 00:18:31.371 Critical Temperature Time: 0 minutes 00:18:31.371 00:18:31.371 Number of Queues 00:18:31.371 ================ 00:18:31.371 Number of I/O Submission Queues: 127 00:18:31.371 Number of I/O Completion Queues: 127 00:18:31.371 00:18:31.371 Active Namespaces 00:18:31.371 ================= 00:18:31.371 Namespace ID:1 00:18:31.371 Error Recovery Timeout: Unlimited 00:18:31.371 Command Set Identifier: NVM (00h) 00:18:31.371 Deallocate: Supported 00:18:31.371 Deallocated/Unwritten Error: Not Supported 00:18:31.371 Deallocated Read Value: Unknown 00:18:31.371 Deallocate in Write Zeroes: Not Supported 00:18:31.371 Deallocated Guard Field: 0xFFFF 00:18:31.371 Flush: Supported 00:18:31.371 Reservation: Supported 00:18:31.371 Namespace Sharing Capabilities: Multiple Controllers 00:18:31.371 Size (in LBAs): 131072 (0GiB) 00:18:31.371 Capacity (in LBAs): 131072 (0GiB) 00:18:31.371 Utilization (in LBAs): 131072 (0GiB) 00:18:31.371 NGUID: 54BBC4E87F6A4516AAA7BA90653444FE 00:18:31.371 UUID: 54bbc4e8-7f6a-4516-aaa7-ba90653444fe 00:18:31.371 Thin Provisioning: Not Supported 00:18:31.371 Per-NS Atomic Units: Yes 00:18:31.371 Atomic Boundary Size (Normal): 0 00:18:31.371 Atomic Boundary Size (PFail): 0 00:18:31.371 Atomic Boundary Offset: 0 00:18:31.371 Maximum Single Source Range Length: 65535 00:18:31.371 Maximum Copy Length: 65535 00:18:31.371 Maximum Source Range Count: 1 00:18:31.371 NGUID/EUI64 Never Reused: No 00:18:31.371 Namespace Write Protected: No 00:18:31.371 Number of LBA Formats: 1 00:18:31.371 Current LBA Format: LBA Format #00 00:18:31.371 LBA Format #00: Data Size: 512 Metadata Size: 0 00:18:31.371 00:18:31.371 08:54:21 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:18:31.371 [2024-11-06 08:54:21.464395] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:18:36.659 Initializing NVMe Controllers 00:18:36.659 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:18:36.659 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:18:36.659 Initialization complete. Launching workers. 00:18:36.659 ======================================================== 00:18:36.660 Latency(us) 00:18:36.660 Device Information : IOPS MiB/s Average min max 00:18:36.660 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 39960.27 156.09 3203.05 846.04 8949.34 00:18:36.660 ======================================================== 00:18:36.660 Total : 39960.27 156.09 3203.05 846.04 8949.34 00:18:36.660 00:18:36.660 [2024-11-06 08:54:26.485140] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:18:36.660 08:54:26 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:18:36.660 [2024-11-06 08:54:26.678003] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:18:41.958 Initializing NVMe Controllers 00:18:41.958 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:18:41.958 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:18:41.958 Initialization complete. Launching workers. 00:18:41.958 ======================================================== 00:18:41.958 Latency(us) 00:18:41.958 Device Information : IOPS MiB/s Average min max 00:18:41.958 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 16051.20 62.70 7980.72 7626.36 8069.01 00:18:41.958 ======================================================== 00:18:41.958 Total : 16051.20 62.70 7980.72 7626.36 8069.01 00:18:41.958 00:18:41.958 [2024-11-06 08:54:31.713050] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:18:41.958 08:54:31 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:18:41.958 [2024-11-06 08:54:31.914972] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:18:47.330 [2024-11-06 08:54:36.987938] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:18:47.330 Initializing NVMe Controllers 00:18:47.330 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:18:47.330 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:18:47.330 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 1 00:18:47.330 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 2 00:18:47.330 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 3 00:18:47.330 Initialization complete. Launching workers. 00:18:47.330 Starting thread on core 2 00:18:47.330 Starting thread on core 3 00:18:47.330 Starting thread on core 1 00:18:47.330 08:54:37 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -d 256 -g 00:18:47.330 [2024-11-06 08:54:37.267723] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:18:51.538 [2024-11-06 08:54:41.055870] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:18:51.538 Initializing NVMe Controllers 00:18:51.538 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:18:51.538 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:18:51.538 Associating SPDK bdev Controller (SPDK1 ) with lcore 0 00:18:51.538 Associating SPDK bdev Controller (SPDK1 ) with lcore 1 00:18:51.538 Associating SPDK bdev Controller (SPDK1 ) with lcore 2 00:18:51.538 Associating SPDK bdev Controller (SPDK1 ) with lcore 3 00:18:51.538 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:18:51.538 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:18:51.538 Initialization complete. Launching workers. 00:18:51.538 Starting thread on core 1 with urgent priority queue 00:18:51.538 Starting thread on core 2 with urgent priority queue 00:18:51.538 Starting thread on core 3 with urgent priority queue 00:18:51.538 Starting thread on core 0 with urgent priority queue 00:18:51.538 SPDK bdev Controller (SPDK1 ) core 0: 14659.00 IO/s 6.82 secs/100000 ios 00:18:51.538 SPDK bdev Controller (SPDK1 ) core 1: 9797.00 IO/s 10.21 secs/100000 ios 00:18:51.538 SPDK bdev Controller (SPDK1 ) core 2: 12575.33 IO/s 7.95 secs/100000 ios 00:18:51.538 SPDK bdev Controller (SPDK1 ) core 3: 8541.67 IO/s 11.71 secs/100000 ios 00:18:51.538 ======================================================== 00:18:51.538 00:18:51.538 08:54:41 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:18:51.538 [2024-11-06 08:54:41.345196] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:18:51.538 Initializing NVMe Controllers 00:18:51.538 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:18:51.538 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:18:51.538 Namespace ID: 1 size: 0GB 00:18:51.538 Initialization complete. 00:18:51.538 INFO: using host memory buffer for IO 00:18:51.538 Hello world! 00:18:51.538 [2024-11-06 08:54:41.381418] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:18:51.538 08:54:41 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:18:51.797 [2024-11-06 08:54:41.669151] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:18:52.742 Initializing NVMe Controllers 00:18:52.742 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:18:52.742 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:18:52.742 Initialization complete. Launching workers. 00:18:52.742 submit (in ns) avg, min, max = 8173.8, 3947.5, 4007044.2 00:18:52.742 complete (in ns) avg, min, max = 18939.7, 2387.5, 4005834.2 00:18:52.742 00:18:52.742 Submit histogram 00:18:52.742 ================ 00:18:52.742 Range in us Cumulative Count 00:18:52.742 3.947 - 3.973: 1.6706% ( 312) 00:18:52.742 3.973 - 4.000: 8.3476% ( 1247) 00:18:52.742 4.000 - 4.027: 18.1516% ( 1831) 00:18:52.742 4.027 - 4.053: 29.8779% ( 2190) 00:18:52.742 4.053 - 4.080: 40.5868% ( 2000) 00:18:52.742 4.080 - 4.107: 52.9342% ( 2306) 00:18:52.742 4.107 - 4.133: 69.5117% ( 3096) 00:18:52.742 4.133 - 4.160: 83.5082% ( 2614) 00:18:52.742 4.160 - 4.187: 92.2200% ( 1627) 00:18:52.742 4.187 - 4.213: 96.8676% ( 868) 00:18:52.742 4.213 - 4.240: 98.5168% ( 308) 00:18:52.742 4.240 - 4.267: 99.2129% ( 130) 00:18:52.742 4.267 - 4.293: 99.4378% ( 42) 00:18:52.742 4.293 - 4.320: 99.4806% ( 8) 00:18:52.742 4.320 - 4.347: 99.4913% ( 2) 00:18:52.742 4.347 - 4.373: 99.5020% ( 2) 00:18:52.742 4.560 - 4.587: 99.5074% ( 1) 00:18:52.742 4.747 - 4.773: 99.5127% ( 1) 00:18:52.742 4.960 - 4.987: 99.5181% ( 1) 00:18:52.742 5.067 - 5.093: 99.5235% ( 1) 00:18:52.742 5.333 - 5.360: 99.5288% ( 1) 00:18:52.742 5.547 - 5.573: 99.5342% ( 1) 00:18:52.742 5.707 - 5.733: 99.5395% ( 1) 00:18:52.742 5.973 - 6.000: 99.5502% ( 2) 00:18:52.742 6.000 - 6.027: 99.5556% ( 1) 00:18:52.742 6.080 - 6.107: 99.5609% ( 1) 00:18:52.742 6.133 - 6.160: 99.5663% ( 1) 00:18:52.742 6.160 - 6.187: 99.5716% ( 1) 00:18:52.742 6.187 - 6.213: 99.5770% ( 1) 00:18:52.742 6.213 - 6.240: 99.5877% ( 2) 00:18:52.742 6.240 - 6.267: 99.6038% ( 3) 00:18:52.742 6.320 - 6.347: 99.6091% ( 1) 00:18:52.742 6.347 - 6.373: 99.6145% ( 1) 00:18:52.742 6.373 - 6.400: 99.6198% ( 1) 00:18:52.742 6.453 - 6.480: 99.6252% ( 1) 00:18:52.742 6.560 - 6.587: 99.6305% ( 1) 00:18:52.742 6.587 - 6.613: 99.6359% ( 1) 00:18:52.742 6.640 - 6.667: 99.6413% ( 1) 00:18:52.742 6.667 - 6.693: 99.6466% ( 1) 00:18:52.742 6.747 - 6.773: 99.6520% ( 1) 00:18:52.742 6.773 - 6.800: 99.6627% ( 2) 00:18:52.742 6.800 - 6.827: 99.6841% ( 4) 00:18:52.742 6.827 - 6.880: 99.6894% ( 1) 00:18:52.742 6.880 - 6.933: 99.6948% ( 1) 00:18:52.742 7.253 - 7.307: 99.7001% ( 1) 00:18:52.742 7.520 - 7.573: 99.7055% ( 1) 00:18:52.742 7.573 - 7.627: 99.7109% ( 1) 00:18:52.742 7.627 - 7.680: 99.7269% ( 3) 00:18:52.742 7.680 - 7.733: 99.7376% ( 2) 00:18:52.742 7.787 - 7.840: 99.7430% ( 1) 00:18:52.742 7.840 - 7.893: 99.7537% ( 2) 00:18:52.742 7.893 - 7.947: 99.7644% ( 2) 00:18:52.742 7.947 - 8.000: 99.7751% ( 2) 00:18:52.742 8.053 - 8.107: 99.7858% ( 2) 00:18:52.742 8.107 - 8.160: 99.7912% ( 1) 00:18:52.742 8.160 - 8.213: 99.8019% ( 2) 00:18:52.742 8.213 - 8.267: 99.8126% ( 2) 00:18:52.742 8.267 - 8.320: 99.8179% ( 1) 00:18:52.742 8.320 - 8.373: 99.8233% ( 1) 00:18:52.742 8.427 - 8.480: 99.8287% ( 1) 00:18:52.742 8.480 - 8.533: 99.8447% ( 3) 00:18:52.742 8.533 - 8.587: 99.8501% ( 1) 00:18:52.742 8.587 - 8.640: 99.8554% ( 1) 00:18:52.742 8.853 - 8.907: 99.8608% ( 1) 00:18:52.742 8.960 - 9.013: 99.8661% ( 1) 00:18:52.742 9.013 - 9.067: 99.8715% ( 1) 00:18:52.742 9.067 - 9.120: 99.8768% ( 1) 00:18:52.742 9.120 - 9.173: 99.8822% ( 1) 00:18:52.742 9.227 - 9.280: 99.8876% ( 1) 00:18:52.742 11.307 - 11.360: 99.8929% ( 1) 00:18:52.742 45.653 - 45.867: 99.8983% ( 1) 00:18:52.742 3986.773 - 4014.080: 100.0000% ( 19) 00:18:52.742 00:18:52.742 Complete histogram 00:18:52.742 ================== 00:18:52.742 Range in us Cumulative Count 00:18:52.742 2.387 - 2.400: 0.7336% ( 137) 00:18:52.742 2.400 - 2.413: 0.9424% ( 39) 00:18:52.742 2.413 - 2.427: 1.1298% ( 35) 00:18:52.742 2.427 - 2.440: 1.1780% ( 9) 00:18:52.742 2.440 - 2.453: 1.2155% ( 7) 00:18:52.742 2.453 - [2024-11-06 08:54:42.694744] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:18:52.742 2.467: 29.4870% ( 5280) 00:18:52.742 2.467 - 2.480: 47.0497% ( 3280) 00:18:52.742 2.480 - 2.493: 62.6205% ( 2908) 00:18:52.742 2.493 - 2.507: 75.9745% ( 2494) 00:18:52.742 2.507 - 2.520: 79.8726% ( 728) 00:18:52.742 2.520 - 2.533: 82.3517% ( 463) 00:18:52.742 2.533 - 2.547: 88.0221% ( 1059) 00:18:52.742 2.547 - 2.560: 92.8090% ( 894) 00:18:52.742 2.560 - 2.573: 95.9253% ( 582) 00:18:52.742 2.573 - 2.587: 98.0938% ( 405) 00:18:52.742 2.587 - 2.600: 98.9880% ( 167) 00:18:52.742 2.600 - 2.613: 99.2397% ( 47) 00:18:52.742 2.613 - 2.627: 99.2986% ( 11) 00:18:52.742 2.627 - 2.640: 99.3146% ( 3) 00:18:52.742 2.640 - 2.653: 99.3200% ( 1) 00:18:52.742 2.653 - 2.667: 99.3253% ( 1) 00:18:52.742 2.693 - 2.707: 99.3307% ( 1) 00:18:52.742 2.707 - 2.720: 99.3360% ( 1) 00:18:52.742 4.587 - 4.613: 99.3468% ( 2) 00:18:52.742 4.667 - 4.693: 99.3575% ( 2) 00:18:52.742 4.720 - 4.747: 99.3682% ( 2) 00:18:52.742 4.773 - 4.800: 99.3735% ( 1) 00:18:52.742 4.853 - 4.880: 99.3789% ( 1) 00:18:52.742 4.907 - 4.933: 99.3842% ( 1) 00:18:52.742 4.933 - 4.960: 99.3896% ( 1) 00:18:52.742 5.040 - 5.067: 99.3949% ( 1) 00:18:52.742 5.093 - 5.120: 99.4003% ( 1) 00:18:52.742 5.120 - 5.147: 99.4057% ( 1) 00:18:52.742 5.147 - 5.173: 99.4110% ( 1) 00:18:52.742 5.200 - 5.227: 99.4164% ( 1) 00:18:52.742 5.227 - 5.253: 99.4217% ( 1) 00:18:52.742 5.253 - 5.280: 99.4271% ( 1) 00:18:52.742 5.307 - 5.333: 99.4324% ( 1) 00:18:52.742 5.360 - 5.387: 99.4431% ( 2) 00:18:52.743 5.440 - 5.467: 99.4538% ( 2) 00:18:52.743 5.547 - 5.573: 99.4592% ( 1) 00:18:52.743 5.733 - 5.760: 99.4646% ( 1) 00:18:52.743 5.867 - 5.893: 99.4699% ( 1) 00:18:52.743 5.893 - 5.920: 99.4753% ( 1) 00:18:52.743 5.920 - 5.947: 99.4860% ( 2) 00:18:52.743 6.027 - 6.053: 99.4913% ( 1) 00:18:52.743 6.107 - 6.133: 99.4967% ( 1) 00:18:52.743 6.160 - 6.187: 99.5020% ( 1) 00:18:52.743 6.213 - 6.240: 99.5074% ( 1) 00:18:52.743 6.320 - 6.347: 99.5127% ( 1) 00:18:52.743 6.373 - 6.400: 99.5181% ( 1) 00:18:52.743 6.667 - 6.693: 99.5235% ( 1) 00:18:52.743 6.747 - 6.773: 99.5288% ( 1) 00:18:52.743 6.800 - 6.827: 99.5395% ( 2) 00:18:52.743 6.827 - 6.880: 99.5449% ( 1) 00:18:52.743 6.987 - 7.040: 99.5502% ( 1) 00:18:52.743 7.307 - 7.360: 99.5663% ( 3) 00:18:52.743 7.467 - 7.520: 99.5770% ( 2) 00:18:52.743 7.787 - 7.840: 99.5824% ( 1) 00:18:52.743 8.000 - 8.053: 99.5877% ( 1) 00:18:52.743 3345.067 - 3358.720: 99.5931% ( 1) 00:18:52.743 3986.773 - 4014.080: 100.0000% ( 76) 00:18:52.743 00:18:52.743 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user1/1 nqn.2019-07.io.spdk:cnode1 1 00:18:52.743 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user1/1 00:18:52.743 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode1 00:18:52.743 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc3 00:18:52.743 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:18:53.004 [ 00:18:53.004 { 00:18:53.004 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:18:53.004 "subtype": "Discovery", 00:18:53.004 "listen_addresses": [], 00:18:53.004 "allow_any_host": true, 00:18:53.004 "hosts": [] 00:18:53.004 }, 00:18:53.004 { 00:18:53.004 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:18:53.004 "subtype": "NVMe", 00:18:53.004 "listen_addresses": [ 00:18:53.004 { 00:18:53.004 "trtype": "VFIOUSER", 00:18:53.004 "adrfam": "IPv4", 00:18:53.004 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:18:53.004 "trsvcid": "0" 00:18:53.004 } 00:18:53.004 ], 00:18:53.004 "allow_any_host": true, 00:18:53.004 "hosts": [], 00:18:53.004 "serial_number": "SPDK1", 00:18:53.004 "model_number": "SPDK bdev Controller", 00:18:53.004 "max_namespaces": 32, 00:18:53.004 "min_cntlid": 1, 00:18:53.004 "max_cntlid": 65519, 00:18:53.004 "namespaces": [ 00:18:53.004 { 00:18:53.004 "nsid": 1, 00:18:53.004 "bdev_name": "Malloc1", 00:18:53.004 "name": "Malloc1", 00:18:53.004 "nguid": "54BBC4E87F6A4516AAA7BA90653444FE", 00:18:53.004 "uuid": "54bbc4e8-7f6a-4516-aaa7-ba90653444fe" 00:18:53.004 } 00:18:53.004 ] 00:18:53.004 }, 00:18:53.004 { 00:18:53.004 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:18:53.004 "subtype": "NVMe", 00:18:53.004 "listen_addresses": [ 00:18:53.004 { 00:18:53.004 "trtype": "VFIOUSER", 00:18:53.004 "adrfam": "IPv4", 00:18:53.004 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:18:53.004 "trsvcid": "0" 00:18:53.004 } 00:18:53.004 ], 00:18:53.004 "allow_any_host": true, 00:18:53.004 "hosts": [], 00:18:53.004 "serial_number": "SPDK2", 00:18:53.004 "model_number": "SPDK bdev Controller", 00:18:53.004 "max_namespaces": 32, 00:18:53.004 "min_cntlid": 1, 00:18:53.004 "max_cntlid": 65519, 00:18:53.004 "namespaces": [ 00:18:53.004 { 00:18:53.004 "nsid": 1, 00:18:53.004 "bdev_name": "Malloc2", 00:18:53.004 "name": "Malloc2", 00:18:53.004 "nguid": "20A1EEAD8C8246F7968A609F0FDD5DD9", 00:18:53.004 "uuid": "20a1eead-8c82-46f7-968a-609f0fdd5dd9" 00:18:53.004 } 00:18:53.004 ] 00:18:53.004 } 00:18:53.004 ] 00:18:53.004 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:18:53.004 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@34 -- # aerpid=1849768 00:18:53.004 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:18:53.004 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -n 2 -g -t /tmp/aer_touch_file 00:18:53.004 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1265 -- # local i=0 00:18:53.004 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:18:53.004 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1272 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:18:53.004 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1276 -- # return 0 00:18:53.004 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:18:53.004 08:54:42 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc3 00:18:53.004 Malloc3 00:18:53.265 [2024-11-06 08:54:43.123179] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:18:53.265 08:54:43 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc3 -n 2 00:18:53.265 [2024-11-06 08:54:43.300355] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:18:53.265 08:54:43 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:18:53.265 Asynchronous Event Request test 00:18:53.265 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:18:53.265 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:18:53.265 Registering asynchronous event callbacks... 00:18:53.265 Starting namespace attribute notice tests for all controllers... 00:18:53.265 /var/run/vfio-user/domain/vfio-user1/1: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:18:53.265 aer_cb - Changed Namespace 00:18:53.265 Cleaning up... 00:18:53.527 [ 00:18:53.527 { 00:18:53.527 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:18:53.527 "subtype": "Discovery", 00:18:53.527 "listen_addresses": [], 00:18:53.527 "allow_any_host": true, 00:18:53.527 "hosts": [] 00:18:53.527 }, 00:18:53.527 { 00:18:53.527 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:18:53.527 "subtype": "NVMe", 00:18:53.527 "listen_addresses": [ 00:18:53.527 { 00:18:53.527 "trtype": "VFIOUSER", 00:18:53.527 "adrfam": "IPv4", 00:18:53.527 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:18:53.527 "trsvcid": "0" 00:18:53.527 } 00:18:53.527 ], 00:18:53.527 "allow_any_host": true, 00:18:53.527 "hosts": [], 00:18:53.527 "serial_number": "SPDK1", 00:18:53.527 "model_number": "SPDK bdev Controller", 00:18:53.527 "max_namespaces": 32, 00:18:53.527 "min_cntlid": 1, 00:18:53.527 "max_cntlid": 65519, 00:18:53.527 "namespaces": [ 00:18:53.527 { 00:18:53.527 "nsid": 1, 00:18:53.527 "bdev_name": "Malloc1", 00:18:53.527 "name": "Malloc1", 00:18:53.527 "nguid": "54BBC4E87F6A4516AAA7BA90653444FE", 00:18:53.527 "uuid": "54bbc4e8-7f6a-4516-aaa7-ba90653444fe" 00:18:53.527 }, 00:18:53.527 { 00:18:53.527 "nsid": 2, 00:18:53.527 "bdev_name": "Malloc3", 00:18:53.527 "name": "Malloc3", 00:18:53.527 "nguid": "ECED2EBD19B44D5B9E37D0D3B13DDD63", 00:18:53.527 "uuid": "eced2ebd-19b4-4d5b-9e37-d0d3b13ddd63" 00:18:53.527 } 00:18:53.527 ] 00:18:53.527 }, 00:18:53.527 { 00:18:53.527 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:18:53.527 "subtype": "NVMe", 00:18:53.527 "listen_addresses": [ 00:18:53.527 { 00:18:53.527 "trtype": "VFIOUSER", 00:18:53.527 "adrfam": "IPv4", 00:18:53.527 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:18:53.527 "trsvcid": "0" 00:18:53.527 } 00:18:53.527 ], 00:18:53.527 "allow_any_host": true, 00:18:53.527 "hosts": [], 00:18:53.527 "serial_number": "SPDK2", 00:18:53.527 "model_number": "SPDK bdev Controller", 00:18:53.527 "max_namespaces": 32, 00:18:53.527 "min_cntlid": 1, 00:18:53.527 "max_cntlid": 65519, 00:18:53.527 "namespaces": [ 00:18:53.527 { 00:18:53.527 "nsid": 1, 00:18:53.527 "bdev_name": "Malloc2", 00:18:53.527 "name": "Malloc2", 00:18:53.527 "nguid": "20A1EEAD8C8246F7968A609F0FDD5DD9", 00:18:53.527 "uuid": "20a1eead-8c82-46f7-968a-609f0fdd5dd9" 00:18:53.527 } 00:18:53.527 ] 00:18:53.527 } 00:18:53.527 ] 00:18:53.527 08:54:43 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@44 -- # wait 1849768 00:18:53.527 08:54:43 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:18:53.527 08:54:43 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user2/2 00:18:53.527 08:54:43 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode2 00:18:53.527 08:54:43 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -L nvme -L nvme_vfio -L vfio_pci 00:18:53.527 [2024-11-06 08:54:43.525635] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:18:53.527 [2024-11-06 08:54:43.525682] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1849822 ] 00:18:53.527 [2024-11-06 08:54:43.577459] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user2/2 00:18:53.527 [2024-11-06 08:54:43.589991] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:18:53.527 [2024-11-06 08:54:43.590013] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7fc194250000 00:18:53.527 [2024-11-06 08:54:43.590993] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:18:53.527 [2024-11-06 08:54:43.591999] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:18:53.527 [2024-11-06 08:54:43.593010] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:18:53.527 [2024-11-06 08:54:43.594018] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:18:53.527 [2024-11-06 08:54:43.595024] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:18:53.527 [2024-11-06 08:54:43.596030] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:18:53.527 [2024-11-06 08:54:43.597033] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:18:53.527 [2024-11-06 08:54:43.598046] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:18:53.527 [2024-11-06 08:54:43.599060] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:18:53.527 [2024-11-06 08:54:43.599074] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7fc194245000 00:18:53.527 [2024-11-06 08:54:43.600399] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:18:53.527 [2024-11-06 08:54:43.616607] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user2/2/cntrl Setup Successfully 00:18:53.528 [2024-11-06 08:54:43.616633] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to connect adminq (no timeout) 00:18:53.528 [2024-11-06 08:54:43.618697] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:18:53.528 [2024-11-06 08:54:43.618744] nvme_pcie_common.c: 134:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:18:53.528 [2024-11-06 08:54:43.618833] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for connect adminq (no timeout) 00:18:53.528 [2024-11-06 08:54:43.618848] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read vs (no timeout) 00:18:53.528 [2024-11-06 08:54:43.618853] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read vs wait for vs (no timeout) 00:18:53.528 [2024-11-06 08:54:43.620755] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x8, value 0x10300 00:18:53.528 [2024-11-06 08:54:43.620766] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read cap (no timeout) 00:18:53.528 [2024-11-06 08:54:43.620773] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to read cap wait for cap (no timeout) 00:18:53.528 [2024-11-06 08:54:43.621718] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:18:53.528 [2024-11-06 08:54:43.621726] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to check en (no timeout) 00:18:53.528 [2024-11-06 08:54:43.621734] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to check en wait for cc (timeout 15000 ms) 00:18:53.528 [2024-11-06 08:54:43.622726] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x0 00:18:53.528 [2024-11-06 08:54:43.622735] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:18:53.528 [2024-11-06 08:54:43.623737] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x0 00:18:53.528 [2024-11-06 08:54:43.623749] nvme_ctrlr.c:3870:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] CC.EN = 0 && CSTS.RDY = 0 00:18:53.528 [2024-11-06 08:54:43.623755] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to controller is disabled (timeout 15000 ms) 00:18:53.528 [2024-11-06 08:54:43.623762] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:18:53.528 [2024-11-06 08:54:43.623867] nvme_ctrlr.c:4068:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Setting CC.EN = 1 00:18:53.528 [2024-11-06 08:54:43.623872] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:18:53.528 [2024-11-06 08:54:43.623877] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x28, value 0x2000003c0000 00:18:53.528 [2024-11-06 08:54:43.624741] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x30, value 0x2000003be000 00:18:53.528 [2024-11-06 08:54:43.625743] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x24, value 0xff00ff 00:18:53.528 [2024-11-06 08:54:43.626756] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:18:53.528 [2024-11-06 08:54:43.627760] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:18:53.528 [2024-11-06 08:54:43.627802] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:18:53.528 [2024-11-06 08:54:43.628770] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x1 00:18:53.528 [2024-11-06 08:54:43.628780] nvme_ctrlr.c:3905:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:18:53.528 [2024-11-06 08:54:43.628785] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to reset admin queue (timeout 30000 ms) 00:18:53.528 [2024-11-06 08:54:43.628807] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify controller (no timeout) 00:18:53.528 [2024-11-06 08:54:43.628815] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify controller (timeout 30000 ms) 00:18:53.528 [2024-11-06 08:54:43.628827] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:18:53.528 [2024-11-06 08:54:43.628832] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:18:53.528 [2024-11-06 08:54:43.628836] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:53.528 [2024-11-06 08:54:43.628848] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:18:53.790 [2024-11-06 08:54:43.639756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:18:53.790 [2024-11-06 08:54:43.639769] nvme_ctrlr.c:2054:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] transport max_xfer_size 131072 00:18:53.790 [2024-11-06 08:54:43.639774] nvme_ctrlr.c:2058:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] MDTS max_xfer_size 131072 00:18:53.790 [2024-11-06 08:54:43.639779] nvme_ctrlr.c:2061:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] CNTLID 0x0001 00:18:53.790 [2024-11-06 08:54:43.639784] nvme_ctrlr.c:2072:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:18:53.790 [2024-11-06 08:54:43.639789] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] transport max_sges 1 00:18:53.790 [2024-11-06 08:54:43.639793] nvme_ctrlr.c:2100:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] fuses compare and write: 1 00:18:53.791 [2024-11-06 08:54:43.639798] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to configure AER (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.639806] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for configure aer (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.639816] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.647752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.647768] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:18:53.791 [2024-11-06 08:54:43.647777] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:18:53.791 [2024-11-06 08:54:43.647786] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:18:53.791 [2024-11-06 08:54:43.647794] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:18:53.791 [2024-11-06 08:54:43.647799] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set keep alive timeout (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.647809] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.647818] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.655752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.655762] nvme_ctrlr.c:3011:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Controller adjusted keep alive timeout to 0 ms 00:18:53.791 [2024-11-06 08:54:43.655767] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify controller iocs specific (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.655774] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set number of queues (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.655780] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for set number of queues (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.655789] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.663750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.663815] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify active ns (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.663824] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify active ns (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.663831] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:18:53.791 [2024-11-06 08:54:43.663836] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:18:53.791 [2024-11-06 08:54:43.663840] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:53.791 [2024-11-06 08:54:43.663846] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.671752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.671763] nvme_ctrlr.c:4699:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Namespace 1 was added 00:18:53.791 [2024-11-06 08:54:43.671775] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify ns (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.671783] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify ns (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.671790] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:18:53.791 [2024-11-06 08:54:43.671795] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:18:53.791 [2024-11-06 08:54:43.671799] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:53.791 [2024-11-06 08:54:43.671805] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.679751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.679768] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify namespace id descriptors (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.679778] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.679785] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:18:53.791 [2024-11-06 08:54:43.679790] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:18:53.791 [2024-11-06 08:54:43.679793] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:53.791 [2024-11-06 08:54:43.679800] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.687752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.687762] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to identify ns iocs specific (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.687769] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set supported log pages (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.687777] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set supported features (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.687783] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set host behavior support feature (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.687788] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set doorbell buffer config (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.687794] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to set host ID (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.687799] nvme_ctrlr.c:3111:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] NVMe-oF transport - not sending Set Features - Host ID 00:18:53.791 [2024-11-06 08:54:43.687803] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to transport ready (timeout 30000 ms) 00:18:53.791 [2024-11-06 08:54:43.687808] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] setting state to ready (no timeout) 00:18:53.791 [2024-11-06 08:54:43.687825] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.695752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.695766] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.703753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.703766] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.711752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.711766] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.719752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.719768] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:18:53.791 [2024-11-06 08:54:43.719773] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:18:53.791 [2024-11-06 08:54:43.719777] nvme_pcie_common.c:1241:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:18:53.791 [2024-11-06 08:54:43.719783] nvme_pcie_common.c:1257:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:18:53.791 [2024-11-06 08:54:43.719786] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 2 00:18:53.791 [2024-11-06 08:54:43.719792] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:18:53.791 [2024-11-06 08:54:43.719800] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:18:53.791 [2024-11-06 08:54:43.719805] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:18:53.791 [2024-11-06 08:54:43.719808] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:53.791 [2024-11-06 08:54:43.719814] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.719822] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:18:53.791 [2024-11-06 08:54:43.719826] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:18:53.791 [2024-11-06 08:54:43.719830] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:53.791 [2024-11-06 08:54:43.719835] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.719845] nvme_pcie_common.c:1204:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:18:53.791 [2024-11-06 08:54:43.719850] nvme_pcie_common.c:1232:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:18:53.791 [2024-11-06 08:54:43.719853] nvme_pcie_common.c:1292:nvme_pcie_qpair_build_contig_request: *DEBUG*: Number of PRP entries: 1 00:18:53.791 [2024-11-06 08:54:43.719859] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:18:53.791 [2024-11-06 08:54:43.727753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.727768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.727778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:18:53.791 [2024-11-06 08:54:43.727786] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:18:53.791 ===================================================== 00:18:53.791 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:18:53.791 ===================================================== 00:18:53.791 Controller Capabilities/Features 00:18:53.792 ================================ 00:18:53.792 Vendor ID: 4e58 00:18:53.792 Subsystem Vendor ID: 4e58 00:18:53.792 Serial Number: SPDK2 00:18:53.792 Model Number: SPDK bdev Controller 00:18:53.792 Firmware Version: 25.01 00:18:53.792 Recommended Arb Burst: 6 00:18:53.792 IEEE OUI Identifier: 8d 6b 50 00:18:53.792 Multi-path I/O 00:18:53.792 May have multiple subsystem ports: Yes 00:18:53.792 May have multiple controllers: Yes 00:18:53.792 Associated with SR-IOV VF: No 00:18:53.792 Max Data Transfer Size: 131072 00:18:53.792 Max Number of Namespaces: 32 00:18:53.792 Max Number of I/O Queues: 127 00:18:53.792 NVMe Specification Version (VS): 1.3 00:18:53.792 NVMe Specification Version (Identify): 1.3 00:18:53.792 Maximum Queue Entries: 256 00:18:53.792 Contiguous Queues Required: Yes 00:18:53.792 Arbitration Mechanisms Supported 00:18:53.792 Weighted Round Robin: Not Supported 00:18:53.792 Vendor Specific: Not Supported 00:18:53.792 Reset Timeout: 15000 ms 00:18:53.792 Doorbell Stride: 4 bytes 00:18:53.792 NVM Subsystem Reset: Not Supported 00:18:53.792 Command Sets Supported 00:18:53.792 NVM Command Set: Supported 00:18:53.792 Boot Partition: Not Supported 00:18:53.792 Memory Page Size Minimum: 4096 bytes 00:18:53.792 Memory Page Size Maximum: 4096 bytes 00:18:53.792 Persistent Memory Region: Not Supported 00:18:53.792 Optional Asynchronous Events Supported 00:18:53.792 Namespace Attribute Notices: Supported 00:18:53.792 Firmware Activation Notices: Not Supported 00:18:53.792 ANA Change Notices: Not Supported 00:18:53.792 PLE Aggregate Log Change Notices: Not Supported 00:18:53.792 LBA Status Info Alert Notices: Not Supported 00:18:53.792 EGE Aggregate Log Change Notices: Not Supported 00:18:53.792 Normal NVM Subsystem Shutdown event: Not Supported 00:18:53.792 Zone Descriptor Change Notices: Not Supported 00:18:53.792 Discovery Log Change Notices: Not Supported 00:18:53.792 Controller Attributes 00:18:53.792 128-bit Host Identifier: Supported 00:18:53.792 Non-Operational Permissive Mode: Not Supported 00:18:53.792 NVM Sets: Not Supported 00:18:53.792 Read Recovery Levels: Not Supported 00:18:53.792 Endurance Groups: Not Supported 00:18:53.792 Predictable Latency Mode: Not Supported 00:18:53.792 Traffic Based Keep ALive: Not Supported 00:18:53.792 Namespace Granularity: Not Supported 00:18:53.792 SQ Associations: Not Supported 00:18:53.792 UUID List: Not Supported 00:18:53.792 Multi-Domain Subsystem: Not Supported 00:18:53.792 Fixed Capacity Management: Not Supported 00:18:53.792 Variable Capacity Management: Not Supported 00:18:53.792 Delete Endurance Group: Not Supported 00:18:53.792 Delete NVM Set: Not Supported 00:18:53.792 Extended LBA Formats Supported: Not Supported 00:18:53.792 Flexible Data Placement Supported: Not Supported 00:18:53.792 00:18:53.792 Controller Memory Buffer Support 00:18:53.792 ================================ 00:18:53.792 Supported: No 00:18:53.792 00:18:53.792 Persistent Memory Region Support 00:18:53.792 ================================ 00:18:53.792 Supported: No 00:18:53.792 00:18:53.792 Admin Command Set Attributes 00:18:53.792 ============================ 00:18:53.792 Security Send/Receive: Not Supported 00:18:53.792 Format NVM: Not Supported 00:18:53.792 Firmware Activate/Download: Not Supported 00:18:53.792 Namespace Management: Not Supported 00:18:53.792 Device Self-Test: Not Supported 00:18:53.792 Directives: Not Supported 00:18:53.792 NVMe-MI: Not Supported 00:18:53.792 Virtualization Management: Not Supported 00:18:53.792 Doorbell Buffer Config: Not Supported 00:18:53.792 Get LBA Status Capability: Not Supported 00:18:53.792 Command & Feature Lockdown Capability: Not Supported 00:18:53.792 Abort Command Limit: 4 00:18:53.792 Async Event Request Limit: 4 00:18:53.792 Number of Firmware Slots: N/A 00:18:53.792 Firmware Slot 1 Read-Only: N/A 00:18:53.792 Firmware Activation Without Reset: N/A 00:18:53.792 Multiple Update Detection Support: N/A 00:18:53.792 Firmware Update Granularity: No Information Provided 00:18:53.792 Per-Namespace SMART Log: No 00:18:53.792 Asymmetric Namespace Access Log Page: Not Supported 00:18:53.792 Subsystem NQN: nqn.2019-07.io.spdk:cnode2 00:18:53.792 Command Effects Log Page: Supported 00:18:53.792 Get Log Page Extended Data: Supported 00:18:53.792 Telemetry Log Pages: Not Supported 00:18:53.792 Persistent Event Log Pages: Not Supported 00:18:53.792 Supported Log Pages Log Page: May Support 00:18:53.792 Commands Supported & Effects Log Page: Not Supported 00:18:53.792 Feature Identifiers & Effects Log Page:May Support 00:18:53.792 NVMe-MI Commands & Effects Log Page: May Support 00:18:53.792 Data Area 4 for Telemetry Log: Not Supported 00:18:53.792 Error Log Page Entries Supported: 128 00:18:53.792 Keep Alive: Supported 00:18:53.792 Keep Alive Granularity: 10000 ms 00:18:53.792 00:18:53.792 NVM Command Set Attributes 00:18:53.792 ========================== 00:18:53.792 Submission Queue Entry Size 00:18:53.792 Max: 64 00:18:53.792 Min: 64 00:18:53.792 Completion Queue Entry Size 00:18:53.792 Max: 16 00:18:53.792 Min: 16 00:18:53.792 Number of Namespaces: 32 00:18:53.792 Compare Command: Supported 00:18:53.792 Write Uncorrectable Command: Not Supported 00:18:53.792 Dataset Management Command: Supported 00:18:53.792 Write Zeroes Command: Supported 00:18:53.792 Set Features Save Field: Not Supported 00:18:53.792 Reservations: Not Supported 00:18:53.792 Timestamp: Not Supported 00:18:53.792 Copy: Supported 00:18:53.792 Volatile Write Cache: Present 00:18:53.792 Atomic Write Unit (Normal): 1 00:18:53.792 Atomic Write Unit (PFail): 1 00:18:53.792 Atomic Compare & Write Unit: 1 00:18:53.792 Fused Compare & Write: Supported 00:18:53.792 Scatter-Gather List 00:18:53.792 SGL Command Set: Supported (Dword aligned) 00:18:53.792 SGL Keyed: Not Supported 00:18:53.792 SGL Bit Bucket Descriptor: Not Supported 00:18:53.792 SGL Metadata Pointer: Not Supported 00:18:53.792 Oversized SGL: Not Supported 00:18:53.792 SGL Metadata Address: Not Supported 00:18:53.792 SGL Offset: Not Supported 00:18:53.792 Transport SGL Data Block: Not Supported 00:18:53.792 Replay Protected Memory Block: Not Supported 00:18:53.792 00:18:53.792 Firmware Slot Information 00:18:53.792 ========================= 00:18:53.792 Active slot: 1 00:18:53.792 Slot 1 Firmware Revision: 25.01 00:18:53.792 00:18:53.792 00:18:53.792 Commands Supported and Effects 00:18:53.792 ============================== 00:18:53.792 Admin Commands 00:18:53.792 -------------- 00:18:53.792 Get Log Page (02h): Supported 00:18:53.792 Identify (06h): Supported 00:18:53.792 Abort (08h): Supported 00:18:53.792 Set Features (09h): Supported 00:18:53.792 Get Features (0Ah): Supported 00:18:53.792 Asynchronous Event Request (0Ch): Supported 00:18:53.792 Keep Alive (18h): Supported 00:18:53.792 I/O Commands 00:18:53.792 ------------ 00:18:53.792 Flush (00h): Supported LBA-Change 00:18:53.792 Write (01h): Supported LBA-Change 00:18:53.792 Read (02h): Supported 00:18:53.792 Compare (05h): Supported 00:18:53.792 Write Zeroes (08h): Supported LBA-Change 00:18:53.792 Dataset Management (09h): Supported LBA-Change 00:18:53.792 Copy (19h): Supported LBA-Change 00:18:53.792 00:18:53.792 Error Log 00:18:53.792 ========= 00:18:53.792 00:18:53.792 Arbitration 00:18:53.792 =========== 00:18:53.792 Arbitration Burst: 1 00:18:53.792 00:18:53.792 Power Management 00:18:53.792 ================ 00:18:53.792 Number of Power States: 1 00:18:53.792 Current Power State: Power State #0 00:18:53.792 Power State #0: 00:18:53.792 Max Power: 0.00 W 00:18:53.792 Non-Operational State: Operational 00:18:53.792 Entry Latency: Not Reported 00:18:53.792 Exit Latency: Not Reported 00:18:53.792 Relative Read Throughput: 0 00:18:53.792 Relative Read Latency: 0 00:18:53.792 Relative Write Throughput: 0 00:18:53.792 Relative Write Latency: 0 00:18:53.792 Idle Power: Not Reported 00:18:53.792 Active Power: Not Reported 00:18:53.792 Non-Operational Permissive Mode: Not Supported 00:18:53.792 00:18:53.792 Health Information 00:18:53.792 ================== 00:18:53.792 Critical Warnings: 00:18:53.792 Available Spare Space: OK 00:18:53.792 Temperature: OK 00:18:53.792 Device Reliability: OK 00:18:53.792 Read Only: No 00:18:53.792 Volatile Memory Backup: OK 00:18:53.792 Current Temperature: 0 Kelvin (-273 Celsius) 00:18:53.792 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:18:53.792 Available Spare: 0% 00:18:53.792 Available Sp[2024-11-06 08:54:43.727886] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:18:53.792 [2024-11-06 08:54:43.735227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:18:53.792 [2024-11-06 08:54:43.735262] nvme_ctrlr.c:4363:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] Prepare to destruct SSD 00:18:53.793 [2024-11-06 08:54:43.735317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:18:53.793 [2024-11-06 08:54:43.735325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:18:53.793 [2024-11-06 08:54:43.735332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:18:53.793 [2024-11-06 08:54:43.735338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:18:53.793 [2024-11-06 08:54:43.735826] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:18:53.793 [2024-11-06 08:54:43.735837] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x464001 00:18:53.793 [2024-11-06 08:54:43.736828] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:18:53.793 [2024-11-06 08:54:43.736878] nvme_ctrlr.c:1124:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] RTD3E = 0 us 00:18:53.793 [2024-11-06 08:54:43.736885] nvme_ctrlr.c:1127:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] shutdown timeout = 10000 ms 00:18:53.793 [2024-11-06 08:54:43.737831] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x9 00:18:53.793 [2024-11-06 08:54:43.737843] nvme_ctrlr.c:1246:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2, 0] shutdown complete in 0 milliseconds 00:18:53.793 [2024-11-06 08:54:43.737891] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user2/2/cntrl 00:18:53.793 [2024-11-06 08:54:43.739270] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:18:53.793 are Threshold: 0% 00:18:53.793 Life Percentage Used: 0% 00:18:53.793 Data Units Read: 0 00:18:53.793 Data Units Written: 0 00:18:53.793 Host Read Commands: 0 00:18:53.793 Host Write Commands: 0 00:18:53.793 Controller Busy Time: 0 minutes 00:18:53.793 Power Cycles: 0 00:18:53.793 Power On Hours: 0 hours 00:18:53.793 Unsafe Shutdowns: 0 00:18:53.793 Unrecoverable Media Errors: 0 00:18:53.793 Lifetime Error Log Entries: 0 00:18:53.793 Warning Temperature Time: 0 minutes 00:18:53.793 Critical Temperature Time: 0 minutes 00:18:53.793 00:18:53.793 Number of Queues 00:18:53.793 ================ 00:18:53.793 Number of I/O Submission Queues: 127 00:18:53.793 Number of I/O Completion Queues: 127 00:18:53.793 00:18:53.793 Active Namespaces 00:18:53.793 ================= 00:18:53.793 Namespace ID:1 00:18:53.793 Error Recovery Timeout: Unlimited 00:18:53.793 Command Set Identifier: NVM (00h) 00:18:53.793 Deallocate: Supported 00:18:53.793 Deallocated/Unwritten Error: Not Supported 00:18:53.793 Deallocated Read Value: Unknown 00:18:53.793 Deallocate in Write Zeroes: Not Supported 00:18:53.793 Deallocated Guard Field: 0xFFFF 00:18:53.793 Flush: Supported 00:18:53.793 Reservation: Supported 00:18:53.793 Namespace Sharing Capabilities: Multiple Controllers 00:18:53.793 Size (in LBAs): 131072 (0GiB) 00:18:53.793 Capacity (in LBAs): 131072 (0GiB) 00:18:53.793 Utilization (in LBAs): 131072 (0GiB) 00:18:53.793 NGUID: 20A1EEAD8C8246F7968A609F0FDD5DD9 00:18:53.793 UUID: 20a1eead-8c82-46f7-968a-609f0fdd5dd9 00:18:53.793 Thin Provisioning: Not Supported 00:18:53.793 Per-NS Atomic Units: Yes 00:18:53.793 Atomic Boundary Size (Normal): 0 00:18:53.793 Atomic Boundary Size (PFail): 0 00:18:53.793 Atomic Boundary Offset: 0 00:18:53.793 Maximum Single Source Range Length: 65535 00:18:53.793 Maximum Copy Length: 65535 00:18:53.793 Maximum Source Range Count: 1 00:18:53.793 NGUID/EUI64 Never Reused: No 00:18:53.793 Namespace Write Protected: No 00:18:53.793 Number of LBA Formats: 1 00:18:53.793 Current LBA Format: LBA Format #00 00:18:53.793 LBA Format #00: Data Size: 512 Metadata Size: 0 00:18:53.793 00:18:53.793 08:54:43 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:18:54.054 [2024-11-06 08:54:43.937139] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:18:59.344 Initializing NVMe Controllers 00:18:59.344 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:18:59.344 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:18:59.344 Initialization complete. Launching workers. 00:18:59.344 ======================================================== 00:18:59.344 Latency(us) 00:18:59.344 Device Information : IOPS MiB/s Average min max 00:18:59.344 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 40033.56 156.38 3197.15 847.74 6983.06 00:18:59.344 ======================================================== 00:18:59.344 Total : 40033.56 156.38 3197.15 847.74 6983.06 00:18:59.344 00:18:59.344 [2024-11-06 08:54:49.040943] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:18:59.344 08:54:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:18:59.344 [2024-11-06 08:54:49.231528] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:19:04.630 Initializing NVMe Controllers 00:19:04.630 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:19:04.630 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:19:04.630 Initialization complete. Launching workers. 00:19:04.630 ======================================================== 00:19:04.630 Latency(us) 00:19:04.630 Device Information : IOPS MiB/s Average min max 00:19:04.630 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 34653.16 135.36 3693.32 1115.20 7481.56 00:19:04.630 ======================================================== 00:19:04.630 Total : 34653.16 135.36 3693.32 1115.20 7481.56 00:19:04.630 00:19:04.630 [2024-11-06 08:54:54.250559] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:19:04.630 08:54:54 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:19:04.630 [2024-11-06 08:54:54.454125] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:19:09.915 [2024-11-06 08:54:59.588834] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:19:09.915 Initializing NVMe Controllers 00:19:09.915 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:19:09.915 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:19:09.915 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 1 00:19:09.915 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 2 00:19:09.915 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 3 00:19:09.915 Initialization complete. Launching workers. 00:19:09.915 Starting thread on core 2 00:19:09.915 Starting thread on core 3 00:19:09.915 Starting thread on core 1 00:19:09.915 08:54:59 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -d 256 -g 00:19:09.915 [2024-11-06 08:54:59.870499] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:19:13.212 [2024-11-06 08:55:02.946530] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:19:13.212 Initializing NVMe Controllers 00:19:13.212 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:19:13.212 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:19:13.212 Associating SPDK bdev Controller (SPDK2 ) with lcore 0 00:19:13.212 Associating SPDK bdev Controller (SPDK2 ) with lcore 1 00:19:13.212 Associating SPDK bdev Controller (SPDK2 ) with lcore 2 00:19:13.212 Associating SPDK bdev Controller (SPDK2 ) with lcore 3 00:19:13.212 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:19:13.212 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:19:13.212 Initialization complete. Launching workers. 00:19:13.212 Starting thread on core 1 with urgent priority queue 00:19:13.212 Starting thread on core 2 with urgent priority queue 00:19:13.212 Starting thread on core 3 with urgent priority queue 00:19:13.212 Starting thread on core 0 with urgent priority queue 00:19:13.212 SPDK bdev Controller (SPDK2 ) core 0: 12726.00 IO/s 7.86 secs/100000 ios 00:19:13.212 SPDK bdev Controller (SPDK2 ) core 1: 8650.67 IO/s 11.56 secs/100000 ios 00:19:13.212 SPDK bdev Controller (SPDK2 ) core 2: 8189.00 IO/s 12.21 secs/100000 ios 00:19:13.212 SPDK bdev Controller (SPDK2 ) core 3: 11750.00 IO/s 8.51 secs/100000 ios 00:19:13.212 ======================================================== 00:19:13.212 00:19:13.212 08:55:03 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:19:13.212 [2024-11-06 08:55:03.236342] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:19:13.212 Initializing NVMe Controllers 00:19:13.212 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:19:13.212 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:19:13.212 Namespace ID: 1 size: 0GB 00:19:13.212 Initialization complete. 00:19:13.212 INFO: using host memory buffer for IO 00:19:13.212 Hello world! 00:19:13.212 [2024-11-06 08:55:03.246400] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:19:13.212 08:55:03 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:19:13.473 [2024-11-06 08:55:03.533738] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:19:14.857 Initializing NVMe Controllers 00:19:14.857 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:19:14.857 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:19:14.857 Initialization complete. Launching workers. 00:19:14.857 submit (in ns) avg, min, max = 8050.9, 3895.0, 4001659.2 00:19:14.857 complete (in ns) avg, min, max = 16560.2, 2409.2, 4000512.5 00:19:14.857 00:19:14.857 Submit histogram 00:19:14.857 ================ 00:19:14.857 Range in us Cumulative Count 00:19:14.857 3.893 - 3.920: 1.8205% ( 346) 00:19:14.857 3.920 - 3.947: 8.4447% ( 1259) 00:19:14.857 3.947 - 3.973: 17.1893% ( 1662) 00:19:14.857 3.973 - 4.000: 28.7541% ( 2198) 00:19:14.857 4.000 - 4.027: 39.5559% ( 2053) 00:19:14.857 4.027 - 4.053: 50.4262% ( 2066) 00:19:14.857 4.053 - 4.080: 65.2320% ( 2814) 00:19:14.857 4.080 - 4.107: 80.8587% ( 2970) 00:19:14.857 4.107 - 4.133: 91.9920% ( 2116) 00:19:14.857 4.133 - 4.160: 97.3114% ( 1011) 00:19:14.857 4.160 - 4.187: 98.9161% ( 305) 00:19:14.857 4.187 - 4.213: 99.4212% ( 96) 00:19:14.857 4.213 - 4.240: 99.4949% ( 14) 00:19:14.857 4.240 - 4.267: 99.5159% ( 4) 00:19:14.857 4.320 - 4.347: 99.5212% ( 1) 00:19:14.857 4.347 - 4.373: 99.5265% ( 1) 00:19:14.857 4.373 - 4.400: 99.5317% ( 1) 00:19:14.857 4.400 - 4.427: 99.5370% ( 1) 00:19:14.857 4.640 - 4.667: 99.5422% ( 1) 00:19:14.857 5.413 - 5.440: 99.5475% ( 1) 00:19:14.857 5.653 - 5.680: 99.5580% ( 2) 00:19:14.857 5.707 - 5.733: 99.5633% ( 1) 00:19:14.857 5.787 - 5.813: 99.5686% ( 1) 00:19:14.857 5.813 - 5.840: 99.5738% ( 1) 00:19:14.857 6.000 - 6.027: 99.5843% ( 2) 00:19:14.857 6.053 - 6.080: 99.5896% ( 1) 00:19:14.858 6.107 - 6.133: 99.5949% ( 1) 00:19:14.858 6.133 - 6.160: 99.6001% ( 1) 00:19:14.858 6.240 - 6.267: 99.6106% ( 2) 00:19:14.858 6.533 - 6.560: 99.6159% ( 1) 00:19:14.858 6.720 - 6.747: 99.6212% ( 1) 00:19:14.858 6.747 - 6.773: 99.6264% ( 1) 00:19:14.858 6.827 - 6.880: 99.6317% ( 1) 00:19:14.858 6.880 - 6.933: 99.6370% ( 1) 00:19:14.858 6.933 - 6.987: 99.6422% ( 1) 00:19:14.858 6.987 - 7.040: 99.6527% ( 2) 00:19:14.858 7.147 - 7.200: 99.6633% ( 2) 00:19:14.858 7.200 - 7.253: 99.6738% ( 2) 00:19:14.858 7.253 - 7.307: 99.6790% ( 1) 00:19:14.858 7.360 - 7.413: 99.7054% ( 5) 00:19:14.858 7.413 - 7.467: 99.7159% ( 2) 00:19:14.858 7.467 - 7.520: 99.7317% ( 3) 00:19:14.858 7.520 - 7.573: 99.7369% ( 1) 00:19:14.858 7.627 - 7.680: 99.7422% ( 1) 00:19:14.858 7.680 - 7.733: 99.7580% ( 3) 00:19:14.858 7.733 - 7.787: 99.7790% ( 4) 00:19:14.858 7.787 - 7.840: 99.7948% ( 3) 00:19:14.858 7.840 - 7.893: 99.8053% ( 2) 00:19:14.858 7.893 - 7.947: 99.8106% ( 1) 00:19:14.858 7.947 - 8.000: 99.8211% ( 2) 00:19:14.858 8.000 - 8.053: 99.8316% ( 2) 00:19:14.858 8.053 - 8.107: 99.8369% ( 1) 00:19:14.858 8.107 - 8.160: 99.8422% ( 1) 00:19:14.858 8.160 - 8.213: 99.8527% ( 2) 00:19:14.858 8.213 - 8.267: 99.8579% ( 1) 00:19:14.858 8.267 - 8.320: 99.8632% ( 1) 00:19:14.858 8.533 - 8.587: 99.8685% ( 1) 00:19:14.858 8.693 - 8.747: 99.8737% ( 1) 00:19:14.858 8.747 - 8.800: 99.8790% ( 1) 00:19:14.858 8.960 - 9.013: 99.8842% ( 1) 00:19:14.858 10.453 - 10.507: 99.8895% ( 1) 00:19:14.858 11.680 - 11.733: 99.8948% ( 1) 00:19:14.858 14.720 - 14.827: 99.9000% ( 1) 00:19:14.858 3986.773 - 4014.080: 100.0000% ( 19) 00:19:14.858 00:19:14.858 Complete histogram 00:19:14.858 ================== 00:19:14.858 Range in us Cumulative Count 00:19:14.858 2.400 - 2.413: 0.0579% ( 11) 00:19:14.858 2.413 - 2.427: 0.8997% ( 160) 00:19:14.858 2.427 - 2.440: 1.0313% ( 25) 00:19:14.858 2.440 - 2.453: 1.1681% ( 26) 00:19:14.858 2.453 - 2.467: 21.8983% ( 3940) 00:19:14.858 2.467 - 2.480: 44.9595% ( 4383) 00:19:14.858 2.480 - 2.493: 63.5010% ( 3524) 00:19:14.858 2.493 - 2.507: 75.8392% ( 2345) 00:19:14.858 2.507 - 2.520: 80.3588% ( 859) 00:19:14.858 2.520 - 2.533: 82.5739% ( 421) 00:19:14.858 2.533 - 2.547: 86.8884% ( 820) 00:19:14.858 2.547 - 2.560: 92.1972% ( 1009) 00:19:14.858 2.560 - 2.573: 95.6908% ( 664) 00:19:14.858 2.573 - [2024-11-06 08:55:04.630422] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:19:14.858 2.587: 97.9954% ( 438) 00:19:14.858 2.587 - 2.600: 99.0371% ( 198) 00:19:14.858 2.600 - 2.613: 99.3160% ( 53) 00:19:14.858 2.613 - 2.627: 99.3844% ( 13) 00:19:14.858 5.307 - 5.333: 99.3897% ( 1) 00:19:14.858 5.413 - 5.440: 99.3949% ( 1) 00:19:14.858 5.467 - 5.493: 99.4002% ( 1) 00:19:14.858 5.573 - 5.600: 99.4055% ( 1) 00:19:14.858 5.627 - 5.653: 99.4107% ( 1) 00:19:14.858 5.653 - 5.680: 99.4160% ( 1) 00:19:14.858 5.707 - 5.733: 99.4212% ( 1) 00:19:14.858 5.760 - 5.787: 99.4318% ( 2) 00:19:14.858 5.787 - 5.813: 99.4370% ( 1) 00:19:14.858 5.813 - 5.840: 99.4423% ( 1) 00:19:14.858 5.840 - 5.867: 99.4475% ( 1) 00:19:14.858 5.920 - 5.947: 99.4528% ( 1) 00:19:14.858 5.947 - 5.973: 99.4686% ( 3) 00:19:14.858 6.000 - 6.027: 99.4791% ( 2) 00:19:14.858 6.053 - 6.080: 99.4949% ( 3) 00:19:14.858 6.080 - 6.107: 99.5002% ( 1) 00:19:14.858 6.133 - 6.160: 99.5054% ( 1) 00:19:14.858 6.213 - 6.240: 99.5107% ( 1) 00:19:14.858 6.240 - 6.267: 99.5159% ( 1) 00:19:14.858 6.373 - 6.400: 99.5212% ( 1) 00:19:14.858 6.480 - 6.507: 99.5265% ( 1) 00:19:14.858 6.507 - 6.533: 99.5370% ( 2) 00:19:14.858 6.560 - 6.587: 99.5475% ( 2) 00:19:14.858 6.613 - 6.640: 99.5528% ( 1) 00:19:14.858 6.640 - 6.667: 99.5580% ( 1) 00:19:14.858 6.667 - 6.693: 99.5633% ( 1) 00:19:14.858 6.693 - 6.720: 99.5686% ( 1) 00:19:14.858 6.933 - 6.987: 99.5738% ( 1) 00:19:14.858 6.987 - 7.040: 99.5791% ( 1) 00:19:14.858 7.040 - 7.093: 99.5843% ( 1) 00:19:14.858 7.253 - 7.307: 99.5896% ( 1) 00:19:14.858 7.413 - 7.467: 99.5949% ( 1) 00:19:14.858 7.520 - 7.573: 99.6001% ( 1) 00:19:14.858 7.573 - 7.627: 99.6054% ( 1) 00:19:14.858 7.627 - 7.680: 99.6106% ( 1) 00:19:14.858 7.893 - 7.947: 99.6159% ( 1) 00:19:14.858 8.160 - 8.213: 99.6212% ( 1) 00:19:14.858 8.213 - 8.267: 99.6264% ( 1) 00:19:14.858 11.360 - 11.413: 99.6317% ( 1) 00:19:14.858 12.427 - 12.480: 99.6370% ( 1) 00:19:14.858 15.040 - 15.147: 99.6422% ( 1) 00:19:14.858 44.160 - 44.373: 99.6475% ( 1) 00:19:14.858 3386.027 - 3399.680: 99.6527% ( 1) 00:19:14.858 3986.773 - 4014.080: 100.0000% ( 66) 00:19:14.858 00:19:14.858 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user2/2 nqn.2019-07.io.spdk:cnode2 2 00:19:14.858 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user2/2 00:19:14.858 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode2 00:19:14.858 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc4 00:19:14.858 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:19:14.858 [ 00:19:14.858 { 00:19:14.858 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:19:14.858 "subtype": "Discovery", 00:19:14.858 "listen_addresses": [], 00:19:14.858 "allow_any_host": true, 00:19:14.858 "hosts": [] 00:19:14.858 }, 00:19:14.858 { 00:19:14.858 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:19:14.858 "subtype": "NVMe", 00:19:14.858 "listen_addresses": [ 00:19:14.858 { 00:19:14.858 "trtype": "VFIOUSER", 00:19:14.858 "adrfam": "IPv4", 00:19:14.858 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:19:14.858 "trsvcid": "0" 00:19:14.858 } 00:19:14.858 ], 00:19:14.858 "allow_any_host": true, 00:19:14.858 "hosts": [], 00:19:14.858 "serial_number": "SPDK1", 00:19:14.858 "model_number": "SPDK bdev Controller", 00:19:14.858 "max_namespaces": 32, 00:19:14.858 "min_cntlid": 1, 00:19:14.858 "max_cntlid": 65519, 00:19:14.858 "namespaces": [ 00:19:14.858 { 00:19:14.858 "nsid": 1, 00:19:14.858 "bdev_name": "Malloc1", 00:19:14.858 "name": "Malloc1", 00:19:14.858 "nguid": "54BBC4E87F6A4516AAA7BA90653444FE", 00:19:14.858 "uuid": "54bbc4e8-7f6a-4516-aaa7-ba90653444fe" 00:19:14.858 }, 00:19:14.858 { 00:19:14.858 "nsid": 2, 00:19:14.858 "bdev_name": "Malloc3", 00:19:14.858 "name": "Malloc3", 00:19:14.858 "nguid": "ECED2EBD19B44D5B9E37D0D3B13DDD63", 00:19:14.858 "uuid": "eced2ebd-19b4-4d5b-9e37-d0d3b13ddd63" 00:19:14.858 } 00:19:14.858 ] 00:19:14.858 }, 00:19:14.858 { 00:19:14.858 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:19:14.858 "subtype": "NVMe", 00:19:14.858 "listen_addresses": [ 00:19:14.858 { 00:19:14.858 "trtype": "VFIOUSER", 00:19:14.858 "adrfam": "IPv4", 00:19:14.858 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:19:14.858 "trsvcid": "0" 00:19:14.858 } 00:19:14.858 ], 00:19:14.858 "allow_any_host": true, 00:19:14.858 "hosts": [], 00:19:14.858 "serial_number": "SPDK2", 00:19:14.858 "model_number": "SPDK bdev Controller", 00:19:14.858 "max_namespaces": 32, 00:19:14.858 "min_cntlid": 1, 00:19:14.858 "max_cntlid": 65519, 00:19:14.858 "namespaces": [ 00:19:14.858 { 00:19:14.858 "nsid": 1, 00:19:14.858 "bdev_name": "Malloc2", 00:19:14.858 "name": "Malloc2", 00:19:14.858 "nguid": "20A1EEAD8C8246F7968A609F0FDD5DD9", 00:19:14.858 "uuid": "20a1eead-8c82-46f7-968a-609f0fdd5dd9" 00:19:14.859 } 00:19:14.859 ] 00:19:14.859 } 00:19:14.859 ] 00:19:14.859 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:19:14.859 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@34 -- # aerpid=1853950 00:19:14.859 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:19:14.859 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -n 2 -g -t /tmp/aer_touch_file 00:19:14.859 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1265 -- # local i=0 00:19:14.859 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:19:14.859 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1272 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:19:14.859 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1276 -- # return 0 00:19:14.859 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:19:14.859 08:55:04 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc4 00:19:15.120 Malloc4 00:19:15.120 [2024-11-06 08:55:05.049641] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:19:15.120 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc4 -n 2 00:19:15.120 [2024-11-06 08:55:05.210754] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:19:15.381 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:19:15.381 Asynchronous Event Request test 00:19:15.382 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:19:15.382 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:19:15.382 Registering asynchronous event callbacks... 00:19:15.382 Starting namespace attribute notice tests for all controllers... 00:19:15.382 /var/run/vfio-user/domain/vfio-user2/2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:19:15.382 aer_cb - Changed Namespace 00:19:15.382 Cleaning up... 00:19:15.382 [ 00:19:15.382 { 00:19:15.382 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:19:15.382 "subtype": "Discovery", 00:19:15.382 "listen_addresses": [], 00:19:15.382 "allow_any_host": true, 00:19:15.382 "hosts": [] 00:19:15.382 }, 00:19:15.382 { 00:19:15.382 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:19:15.382 "subtype": "NVMe", 00:19:15.382 "listen_addresses": [ 00:19:15.382 { 00:19:15.382 "trtype": "VFIOUSER", 00:19:15.382 "adrfam": "IPv4", 00:19:15.382 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:19:15.382 "trsvcid": "0" 00:19:15.382 } 00:19:15.382 ], 00:19:15.382 "allow_any_host": true, 00:19:15.382 "hosts": [], 00:19:15.382 "serial_number": "SPDK1", 00:19:15.382 "model_number": "SPDK bdev Controller", 00:19:15.382 "max_namespaces": 32, 00:19:15.382 "min_cntlid": 1, 00:19:15.382 "max_cntlid": 65519, 00:19:15.382 "namespaces": [ 00:19:15.382 { 00:19:15.382 "nsid": 1, 00:19:15.382 "bdev_name": "Malloc1", 00:19:15.382 "name": "Malloc1", 00:19:15.382 "nguid": "54BBC4E87F6A4516AAA7BA90653444FE", 00:19:15.382 "uuid": "54bbc4e8-7f6a-4516-aaa7-ba90653444fe" 00:19:15.382 }, 00:19:15.382 { 00:19:15.382 "nsid": 2, 00:19:15.382 "bdev_name": "Malloc3", 00:19:15.382 "name": "Malloc3", 00:19:15.382 "nguid": "ECED2EBD19B44D5B9E37D0D3B13DDD63", 00:19:15.382 "uuid": "eced2ebd-19b4-4d5b-9e37-d0d3b13ddd63" 00:19:15.382 } 00:19:15.382 ] 00:19:15.382 }, 00:19:15.382 { 00:19:15.382 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:19:15.382 "subtype": "NVMe", 00:19:15.382 "listen_addresses": [ 00:19:15.382 { 00:19:15.382 "trtype": "VFIOUSER", 00:19:15.382 "adrfam": "IPv4", 00:19:15.382 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:19:15.382 "trsvcid": "0" 00:19:15.382 } 00:19:15.382 ], 00:19:15.382 "allow_any_host": true, 00:19:15.382 "hosts": [], 00:19:15.382 "serial_number": "SPDK2", 00:19:15.382 "model_number": "SPDK bdev Controller", 00:19:15.382 "max_namespaces": 32, 00:19:15.382 "min_cntlid": 1, 00:19:15.382 "max_cntlid": 65519, 00:19:15.382 "namespaces": [ 00:19:15.382 { 00:19:15.382 "nsid": 1, 00:19:15.382 "bdev_name": "Malloc2", 00:19:15.382 "name": "Malloc2", 00:19:15.382 "nguid": "20A1EEAD8C8246F7968A609F0FDD5DD9", 00:19:15.382 "uuid": "20a1eead-8c82-46f7-968a-609f0fdd5dd9" 00:19:15.382 }, 00:19:15.382 { 00:19:15.382 "nsid": 2, 00:19:15.382 "bdev_name": "Malloc4", 00:19:15.382 "name": "Malloc4", 00:19:15.382 "nguid": "B702E9C1B9F344F0B35F9ECA3B5FABC4", 00:19:15.382 "uuid": "b702e9c1-b9f3-44f0-b35f-9eca3b5fabc4" 00:19:15.382 } 00:19:15.382 ] 00:19:15.382 } 00:19:15.382 ] 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@44 -- # wait 1853950 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@105 -- # stop_nvmf_vfio_user 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@95 -- # killprocess 1844749 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@950 -- # '[' -z 1844749 ']' 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@954 -- # kill -0 1844749 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@955 -- # uname 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1844749 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1844749' 00:19:15.382 killing process with pid 1844749 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@969 -- # kill 1844749 00:19:15.382 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@974 -- # wait 1844749 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@108 -- # setup_nvmf_vfio_user --interrupt-mode '-M -I' 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args=--interrupt-mode 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@52 -- # local 'transport_args=-M -I' 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=1854187 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 1854187' 00:19:15.643 Process pid: 1854187 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 1854187 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@831 -- # '[' -z 1854187 ']' 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:15.643 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:19:15.643 08:55:05 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' --interrupt-mode 00:19:15.643 [2024-11-06 08:55:05.704609] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:19:15.643 [2024-11-06 08:55:05.705534] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:19:15.643 [2024-11-06 08:55:05.705578] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:15.904 [2024-11-06 08:55:05.777776] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:19:15.904 [2024-11-06 08:55:05.812410] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:15.904 [2024-11-06 08:55:05.812447] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:15.904 [2024-11-06 08:55:05.812455] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:15.904 [2024-11-06 08:55:05.812461] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:15.904 [2024-11-06 08:55:05.812467] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:15.904 [2024-11-06 08:55:05.813953] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:15.904 [2024-11-06 08:55:05.814123] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:15.904 [2024-11-06 08:55:05.814278] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:19:15.904 [2024-11-06 08:55:05.814279] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:19:15.904 [2024-11-06 08:55:05.870134] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:19:15.904 [2024-11-06 08:55:05.870432] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:19:15.904 [2024-11-06 08:55:05.871336] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:19:15.904 [2024-11-06 08:55:05.871615] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:19:15.904 [2024-11-06 08:55:05.871789] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:19:16.475 08:55:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:16.475 08:55:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@864 -- # return 0 00:19:16.475 08:55:06 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:19:17.419 08:55:07 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER -M -I 00:19:17.681 08:55:07 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:19:17.681 08:55:07 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:19:17.681 08:55:07 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:19:17.681 08:55:07 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:19:17.681 08:55:07 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:19:17.941 Malloc1 00:19:17.941 08:55:07 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:19:18.202 08:55:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:19:18.202 08:55:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:19:18.463 08:55:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:19:18.463 08:55:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:19:18.463 08:55:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:19:18.723 Malloc2 00:19:18.723 08:55:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:19:18.723 08:55:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:19:18.984 08:55:08 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@109 -- # stop_nvmf_vfio_user 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@95 -- # killprocess 1854187 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@950 -- # '[' -z 1854187 ']' 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@954 -- # kill -0 1854187 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@955 -- # uname 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1854187 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1854187' 00:19:19.244 killing process with pid 1854187 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@969 -- # kill 1854187 00:19:19.244 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@974 -- # wait 1854187 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:19:19.505 00:19:19.505 real 0m52.035s 00:19:19.505 user 3m19.601s 00:19:19.505 sys 0m2.696s 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:19:19.505 ************************************ 00:19:19.505 END TEST nvmf_vfio_user 00:19:19.505 ************************************ 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@32 -- # run_test nvmf_vfio_user_nvme_compliance /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:19:19.505 ************************************ 00:19:19.505 START TEST nvmf_vfio_user_nvme_compliance 00:19:19.505 ************************************ 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:19:19.505 * Looking for test storage... 00:19:19.505 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1689 -- # lcov --version 00:19:19.505 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@336 -- # IFS=.-: 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@336 -- # read -ra ver1 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@337 -- # IFS=.-: 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@337 -- # read -ra ver2 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@338 -- # local 'op=<' 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@340 -- # ver1_l=2 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@341 -- # ver2_l=1 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@344 -- # case "$op" in 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@345 -- # : 1 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@365 -- # decimal 1 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@353 -- # local d=1 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@355 -- # echo 1 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@365 -- # ver1[v]=1 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@366 -- # decimal 2 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@353 -- # local d=2 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@355 -- # echo 2 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@366 -- # ver2[v]=2 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@368 -- # return 0 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:19:19.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:19.766 --rc genhtml_branch_coverage=1 00:19:19.766 --rc genhtml_function_coverage=1 00:19:19.766 --rc genhtml_legend=1 00:19:19.766 --rc geninfo_all_blocks=1 00:19:19.766 --rc geninfo_unexecuted_blocks=1 00:19:19.766 00:19:19.766 ' 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:19:19.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:19.766 --rc genhtml_branch_coverage=1 00:19:19.766 --rc genhtml_function_coverage=1 00:19:19.766 --rc genhtml_legend=1 00:19:19.766 --rc geninfo_all_blocks=1 00:19:19.766 --rc geninfo_unexecuted_blocks=1 00:19:19.766 00:19:19.766 ' 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:19:19.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:19.766 --rc genhtml_branch_coverage=1 00:19:19.766 --rc genhtml_function_coverage=1 00:19:19.766 --rc genhtml_legend=1 00:19:19.766 --rc geninfo_all_blocks=1 00:19:19.766 --rc geninfo_unexecuted_blocks=1 00:19:19.766 00:19:19.766 ' 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:19:19.766 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:19.766 --rc genhtml_branch_coverage=1 00:19:19.766 --rc genhtml_function_coverage=1 00:19:19.766 --rc genhtml_legend=1 00:19:19.766 --rc geninfo_all_blocks=1 00:19:19.766 --rc geninfo_unexecuted_blocks=1 00:19:19.766 00:19:19.766 ' 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@7 -- # uname -s 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@15 -- # shopt -s extglob 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@5 -- # export PATH 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@51 -- # : 0 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:19:19.766 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:19:19.767 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@55 -- # have_pci_nics=0 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@11 -- # MALLOC_BDEV_SIZE=64 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@14 -- # export TEST_TRANSPORT=VFIOUSER 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@14 -- # TEST_TRANSPORT=VFIOUSER 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@16 -- # rm -rf /var/run/vfio-user 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@20 -- # nvmfpid=1854946 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@21 -- # echo 'Process pid: 1854946' 00:19:19.767 Process pid: 1854946 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@23 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@24 -- # waitforlisten 1854946 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@831 -- # '[' -z 1854946 ']' 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:19.767 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:19.767 08:55:09 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:19:19.767 [2024-11-06 08:55:09.765463] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:19:19.767 [2024-11-06 08:55:09.765541] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:19.767 [2024-11-06 08:55:09.842240] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:19:20.027 [2024-11-06 08:55:09.883796] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:20.027 [2024-11-06 08:55:09.883829] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:20.027 [2024-11-06 08:55:09.883837] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:20.027 [2024-11-06 08:55:09.883844] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:20.027 [2024-11-06 08:55:09.883849] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:20.027 [2024-11-06 08:55:09.885393] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:19:20.027 [2024-11-06 08:55:09.885533] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:19:20.027 [2024-11-06 08:55:09.885535] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:19:20.598 08:55:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:20.598 08:55:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@864 -- # return 0 00:19:20.598 08:55:10 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@26 -- # sleep 1 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@28 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@29 -- # traddr=/var/run/vfio-user 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@31 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@33 -- # mkdir -p /var/run/vfio-user 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@35 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:19:21.539 malloc0 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@36 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk -m 32 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@37 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@38 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:21.539 08:55:11 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/nvme_compliance -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user subnqn:nqn.2021-09.io.spdk:cnode0' 00:19:21.800 00:19:21.800 00:19:21.800 CUnit - A unit testing framework for C - Version 2.1-3 00:19:21.800 http://cunit.sourceforge.net/ 00:19:21.800 00:19:21.800 00:19:21.800 Suite: nvme_compliance 00:19:21.800 Test: admin_identify_ctrlr_verify_dptr ...[2024-11-06 08:55:11.836198] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:21.800 [2024-11-06 08:55:11.837545] vfio_user.c: 804:nvme_cmd_map_prps: *ERROR*: no PRP2, 3072 remaining 00:19:21.800 [2024-11-06 08:55:11.837558] vfio_user.c:5507:map_admin_cmd_req: *ERROR*: /var/run/vfio-user: map Admin Opc 6 failed 00:19:21.800 [2024-11-06 08:55:11.837562] vfio_user.c:5600:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x6 failed 00:19:21.800 [2024-11-06 08:55:11.839209] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:21.800 passed 00:19:22.061 Test: admin_identify_ctrlr_verify_fused ...[2024-11-06 08:55:11.934785] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:22.061 [2024-11-06 08:55:11.937807] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:22.061 passed 00:19:22.061 Test: admin_identify_ns ...[2024-11-06 08:55:12.033002] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:22.061 [2024-11-06 08:55:12.096759] ctrlr.c:2750:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 0 00:19:22.061 [2024-11-06 08:55:12.104765] ctrlr.c:2750:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 4294967295 00:19:22.061 [2024-11-06 08:55:12.125865] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:22.061 passed 00:19:22.322 Test: admin_get_features_mandatory_features ...[2024-11-06 08:55:12.216502] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:22.322 [2024-11-06 08:55:12.219521] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:22.322 passed 00:19:22.322 Test: admin_get_features_optional_features ...[2024-11-06 08:55:12.314097] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:22.322 [2024-11-06 08:55:12.317114] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:22.322 passed 00:19:22.322 Test: admin_set_features_number_of_queues ...[2024-11-06 08:55:12.410241] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:22.582 [2024-11-06 08:55:12.514853] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:22.582 passed 00:19:22.582 Test: admin_get_log_page_mandatory_logs ...[2024-11-06 08:55:12.612957] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:22.582 [2024-11-06 08:55:12.615980] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:22.582 passed 00:19:22.842 Test: admin_get_log_page_with_lpo ...[2024-11-06 08:55:12.709131] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:22.842 [2024-11-06 08:55:12.776757] ctrlr.c:2697:nvmf_ctrlr_get_log_page: *ERROR*: Get log page: offset (516) > len (512) 00:19:22.842 [2024-11-06 08:55:12.789816] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:22.842 passed 00:19:22.842 Test: fabric_property_get ...[2024-11-06 08:55:12.881415] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:22.842 [2024-11-06 08:55:12.882672] vfio_user.c:5600:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x7f failed 00:19:22.842 [2024-11-06 08:55:12.884441] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:22.842 passed 00:19:23.102 Test: admin_delete_io_sq_use_admin_qid ...[2024-11-06 08:55:12.979070] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:23.102 [2024-11-06 08:55:12.980314] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:0 does not exist 00:19:23.102 [2024-11-06 08:55:12.982090] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:23.102 passed 00:19:23.102 Test: admin_delete_io_sq_delete_sq_twice ...[2024-11-06 08:55:13.073973] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:23.102 [2024-11-06 08:55:13.157754] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:19:23.102 [2024-11-06 08:55:13.173756] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:19:23.102 [2024-11-06 08:55:13.178842] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:23.362 passed 00:19:23.362 Test: admin_delete_io_cq_use_admin_qid ...[2024-11-06 08:55:13.272810] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:23.362 [2024-11-06 08:55:13.274060] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O cqid:0 does not exist 00:19:23.362 [2024-11-06 08:55:13.275829] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:23.362 passed 00:19:23.362 Test: admin_delete_io_cq_delete_cq_first ...[2024-11-06 08:55:13.369006] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:23.362 [2024-11-06 08:55:13.444758] vfio_user.c:2319:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:19:23.362 [2024-11-06 08:55:13.468755] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:19:23.362 [2024-11-06 08:55:13.473841] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:23.622 passed 00:19:23.622 Test: admin_create_io_cq_verify_iv_pc ...[2024-11-06 08:55:13.567831] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:23.622 [2024-11-06 08:55:13.569074] vfio_user.c:2158:handle_create_io_cq: *ERROR*: /var/run/vfio-user: IV is too big 00:19:23.622 [2024-11-06 08:55:13.569095] vfio_user.c:2152:handle_create_io_cq: *ERROR*: /var/run/vfio-user: non-PC CQ not supported 00:19:23.622 [2024-11-06 08:55:13.570853] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:23.622 passed 00:19:23.622 Test: admin_create_io_sq_verify_qsize_cqid ...[2024-11-06 08:55:13.663999] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:23.882 [2024-11-06 08:55:13.755753] vfio_user.c:2240:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 1 00:19:23.882 [2024-11-06 08:55:13.763753] vfio_user.c:2240:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 257 00:19:23.883 [2024-11-06 08:55:13.771755] vfio_user.c:2038:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:0 00:19:23.883 [2024-11-06 08:55:13.779753] vfio_user.c:2038:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:128 00:19:23.883 [2024-11-06 08:55:13.808832] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:23.883 passed 00:19:23.883 Test: admin_create_io_sq_verify_pc ...[2024-11-06 08:55:13.902828] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:23.883 [2024-11-06 08:55:13.922761] vfio_user.c:2051:handle_create_io_sq: *ERROR*: /var/run/vfio-user: non-PC SQ not supported 00:19:23.883 [2024-11-06 08:55:13.940006] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:23.883 passed 00:19:24.143 Test: admin_create_io_qp_max_qps ...[2024-11-06 08:55:14.031550] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:25.082 [2024-11-06 08:55:15.135756] nvme_ctrlr.c:5487:spdk_nvme_ctrlr_alloc_qid: *ERROR*: [/var/run/vfio-user, 0] No free I/O queue IDs 00:19:25.653 [2024-11-06 08:55:15.531015] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:25.653 passed 00:19:25.653 Test: admin_create_io_sq_shared_cq ...[2024-11-06 08:55:15.629202] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:19:25.653 [2024-11-06 08:55:15.761751] vfio_user.c:2319:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:19:25.914 [2024-11-06 08:55:15.798808] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:19:25.914 passed 00:19:25.914 00:19:25.914 Run Summary: Type Total Ran Passed Failed Inactive 00:19:25.914 suites 1 1 n/a 0 0 00:19:25.914 tests 18 18 18 0 0 00:19:25.914 asserts 360 360 360 0 n/a 00:19:25.914 00:19:25.914 Elapsed time = 1.662 seconds 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@42 -- # killprocess 1854946 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@950 -- # '[' -z 1854946 ']' 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@954 -- # kill -0 1854946 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@955 -- # uname 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1854946 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1854946' 00:19:25.914 killing process with pid 1854946 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@969 -- # kill 1854946 00:19:25.914 08:55:15 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@974 -- # wait 1854946 00:19:26.174 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@44 -- # rm -rf /var/run/vfio-user 00:19:26.174 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@46 -- # trap - SIGINT SIGTERM EXIT 00:19:26.174 00:19:26.174 real 0m6.585s 00:19:26.175 user 0m18.676s 00:19:26.175 sys 0m0.551s 00:19:26.175 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1126 -- # xtrace_disable 00:19:26.175 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:19:26.175 ************************************ 00:19:26.175 END TEST nvmf_vfio_user_nvme_compliance 00:19:26.175 ************************************ 00:19:26.175 08:55:16 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@33 -- # run_test nvmf_vfio_user_fuzz /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:19:26.175 08:55:16 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:19:26.175 08:55:16 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:19:26.175 08:55:16 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:19:26.175 ************************************ 00:19:26.175 START TEST nvmf_vfio_user_fuzz 00:19:26.175 ************************************ 00:19:26.175 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:19:26.175 * Looking for test storage... 00:19:26.175 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:19:26.175 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:19:26.175 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1689 -- # lcov --version 00:19:26.175 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@333 -- # local ver1 ver1_l 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@334 -- # local ver2 ver2_l 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@336 -- # IFS=.-: 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@336 -- # read -ra ver1 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@337 -- # IFS=.-: 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@337 -- # read -ra ver2 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@338 -- # local 'op=<' 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@340 -- # ver1_l=2 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@341 -- # ver2_l=1 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@344 -- # case "$op" in 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@345 -- # : 1 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@364 -- # (( v = 0 )) 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@365 -- # decimal 1 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@353 -- # local d=1 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@355 -- # echo 1 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@365 -- # ver1[v]=1 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@366 -- # decimal 2 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@353 -- # local d=2 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@355 -- # echo 2 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@366 -- # ver2[v]=2 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@368 -- # return 0 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:19:26.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:26.436 --rc genhtml_branch_coverage=1 00:19:26.436 --rc genhtml_function_coverage=1 00:19:26.436 --rc genhtml_legend=1 00:19:26.436 --rc geninfo_all_blocks=1 00:19:26.436 --rc geninfo_unexecuted_blocks=1 00:19:26.436 00:19:26.436 ' 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:19:26.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:26.436 --rc genhtml_branch_coverage=1 00:19:26.436 --rc genhtml_function_coverage=1 00:19:26.436 --rc genhtml_legend=1 00:19:26.436 --rc geninfo_all_blocks=1 00:19:26.436 --rc geninfo_unexecuted_blocks=1 00:19:26.436 00:19:26.436 ' 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:19:26.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:26.436 --rc genhtml_branch_coverage=1 00:19:26.436 --rc genhtml_function_coverage=1 00:19:26.436 --rc genhtml_legend=1 00:19:26.436 --rc geninfo_all_blocks=1 00:19:26.436 --rc geninfo_unexecuted_blocks=1 00:19:26.436 00:19:26.436 ' 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:19:26.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:19:26.436 --rc genhtml_branch_coverage=1 00:19:26.436 --rc genhtml_function_coverage=1 00:19:26.436 --rc genhtml_legend=1 00:19:26.436 --rc geninfo_all_blocks=1 00:19:26.436 --rc geninfo_unexecuted_blocks=1 00:19:26.436 00:19:26.436 ' 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@7 -- # uname -s 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:19:26.436 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@15 -- # shopt -s extglob 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@5 -- # export PATH 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@51 -- # : 0 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:19:26.437 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- nvmf/common.sh@55 -- # have_pci_nics=0 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@12 -- # MALLOC_BDEV_SIZE=64 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@15 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@16 -- # traddr=/var/run/vfio-user 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@20 -- # rm -rf /var/run/vfio-user 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@24 -- # nvmfpid=1856345 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@25 -- # echo 'Process pid: 1856345' 00:19:26.437 Process pid: 1856345 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@27 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@28 -- # waitforlisten 1856345 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@831 -- # '[' -z 1856345 ']' 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@836 -- # local max_retries=100 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:26.437 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@840 -- # xtrace_disable 00:19:26.437 08:55:16 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:19:27.379 08:55:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:19:27.379 08:55:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@864 -- # return 0 00:19:27.379 08:55:17 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@30 -- # sleep 1 00:19:28.320 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@32 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:19:28.320 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:28.320 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@34 -- # mkdir -p /var/run/vfio-user 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:19:28.321 malloc0 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@39 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@41 -- # trid='trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' 00:19:28.321 08:55:18 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/fuzz/nvme_fuzz/nvme_fuzz -m 0x2 -t 30 -S 123456 -F 'trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' -N -a 00:20:00.435 Fuzzing completed. Shutting down the fuzz application 00:20:00.435 00:20:00.435 Dumping successful admin opcodes: 00:20:00.435 8, 9, 10, 24, 00:20:00.435 Dumping successful io opcodes: 00:20:00.435 0, 00:20:00.435 NS: 0x20000081ef00 I/O qp, Total commands completed: 1120660, total successful commands: 4411, random_seed: 2000174912 00:20:00.435 NS: 0x20000081ef00 admin qp, Total commands completed: 140908, total successful commands: 1144, random_seed: 3551736128 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@44 -- # rpc_cmd nvmf_delete_subsystem nqn.2021-09.io.spdk:cnode0 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@46 -- # killprocess 1856345 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@950 -- # '[' -z 1856345 ']' 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@954 -- # kill -0 1856345 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@955 -- # uname 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1856345 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1856345' 00:20:00.435 killing process with pid 1856345 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@969 -- # kill 1856345 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@974 -- # wait 1856345 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@48 -- # rm -rf /var/run/vfio-user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_log.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_tgt_output.txt 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@50 -- # trap - SIGINT SIGTERM EXIT 00:20:00.435 00:20:00.435 real 0m33.787s 00:20:00.435 user 0m38.095s 00:20:00.435 sys 0m25.980s 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1126 -- # xtrace_disable 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:20:00.435 ************************************ 00:20:00.435 END TEST nvmf_vfio_user_fuzz 00:20:00.435 ************************************ 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@37 -- # run_test nvmf_auth_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/auth.sh --transport=tcp 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:20:00.435 ************************************ 00:20:00.435 START TEST nvmf_auth_target 00:20:00.435 ************************************ 00:20:00.435 08:55:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/auth.sh --transport=tcp 00:20:00.435 * Looking for test storage... 00:20:00.435 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:20:00.435 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:20:00.435 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1689 -- # lcov --version 00:20:00.435 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:20:00.435 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:20:00.435 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:20:00.435 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:20:00.435 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:20:00.435 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@336 -- # IFS=.-: 00:20:00.435 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@336 -- # read -ra ver1 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@337 -- # IFS=.-: 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@337 -- # read -ra ver2 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@338 -- # local 'op=<' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@340 -- # ver1_l=2 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@341 -- # ver2_l=1 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@344 -- # case "$op" in 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@345 -- # : 1 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@365 -- # decimal 1 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@353 -- # local d=1 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@355 -- # echo 1 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@365 -- # ver1[v]=1 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@366 -- # decimal 2 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@353 -- # local d=2 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@355 -- # echo 2 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@366 -- # ver2[v]=2 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@368 -- # return 0 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:20:00.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:00.436 --rc genhtml_branch_coverage=1 00:20:00.436 --rc genhtml_function_coverage=1 00:20:00.436 --rc genhtml_legend=1 00:20:00.436 --rc geninfo_all_blocks=1 00:20:00.436 --rc geninfo_unexecuted_blocks=1 00:20:00.436 00:20:00.436 ' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:20:00.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:00.436 --rc genhtml_branch_coverage=1 00:20:00.436 --rc genhtml_function_coverage=1 00:20:00.436 --rc genhtml_legend=1 00:20:00.436 --rc geninfo_all_blocks=1 00:20:00.436 --rc geninfo_unexecuted_blocks=1 00:20:00.436 00:20:00.436 ' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:20:00.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:00.436 --rc genhtml_branch_coverage=1 00:20:00.436 --rc genhtml_function_coverage=1 00:20:00.436 --rc genhtml_legend=1 00:20:00.436 --rc geninfo_all_blocks=1 00:20:00.436 --rc geninfo_unexecuted_blocks=1 00:20:00.436 00:20:00.436 ' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:20:00.436 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:20:00.436 --rc genhtml_branch_coverage=1 00:20:00.436 --rc genhtml_function_coverage=1 00:20:00.436 --rc genhtml_legend=1 00:20:00.436 --rc geninfo_all_blocks=1 00:20:00.436 --rc geninfo_unexecuted_blocks=1 00:20:00.436 00:20:00.436 ' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@7 -- # uname -s 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@15 -- # shopt -s extglob 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@5 -- # export PATH 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@51 -- # : 0 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:20:00.436 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@14 -- # dhgroups=("null" "ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@15 -- # subnqn=nqn.2024-03.io.spdk:cnode0 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@16 -- # hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@17 -- # hostsock=/var/tmp/host.sock 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@18 -- # keys=() 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@18 -- # ckeys=() 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@86 -- # nvmftestinit 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:00.436 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@474 -- # prepare_net_devs 00:20:00.437 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@436 -- # local -g is_hw=no 00:20:00.437 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@438 -- # remove_spdk_ns 00:20:00.437 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:00.437 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:20:00.437 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:00.437 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:20:00.437 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:20:00.437 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@309 -- # xtrace_disable 00:20:00.437 08:55:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@315 -- # pci_devs=() 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@319 -- # net_devs=() 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@320 -- # e810=() 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@320 -- # local -ga e810 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@321 -- # x722=() 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@321 -- # local -ga x722 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@322 -- # mlx=() 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@322 -- # local -ga mlx 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:20:08.632 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:20:08.632 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:20:08.632 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:20:08.633 Found net devices under 0000:4b:00.0: cvl_0_0 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:20:08.633 Found net devices under 0000:4b:00.1: cvl_0_1 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@440 -- # is_hw=yes 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:20:08.633 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:08.633 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.635 ms 00:20:08.633 00:20:08.633 --- 10.0.0.2 ping statistics --- 00:20:08.633 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:08.633 rtt min/avg/max/mdev = 0.635/0.635/0.635/0.000 ms 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:08.633 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:08.633 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.280 ms 00:20:08.633 00:20:08.633 --- 10.0.0.1 ping statistics --- 00:20:08.633 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:08.633 rtt min/avg/max/mdev = 0.280/0.280/0.280/0.000 ms 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@448 -- # return 0 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@87 -- # nvmfappstart -L nvmf_auth 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@724 -- # xtrace_disable 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@507 -- # nvmfpid=1866650 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@508 -- # waitforlisten 1866650 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvmf_auth 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@831 -- # '[' -z 1866650 ']' 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # return 0 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@730 -- # xtrace_disable 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@89 -- # hostpid=1866679 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@91 -- # trap 'dumplogs; cleanup' SIGINT SIGTERM EXIT 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 2 -r /var/tmp/host.sock -L nvme_auth 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # gen_dhchap_key null 48 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=null 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=48 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=91c5d9c7ce279e23d21763f686209e69f36237c6ef8527f9 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-null.XXX 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-null.2c1 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key 91c5d9c7ce279e23d21763f686209e69f36237c6ef8527f9 0 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 91c5d9c7ce279e23d21763f686209e69f36237c6ef8527f9 0 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:20:08.633 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=91c5d9c7ce279e23d21763f686209e69f36237c6ef8527f9 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=0 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-null.2c1 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-null.2c1 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # keys[0]=/tmp/spdk.key-null.2c1 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # gen_dhchap_key sha512 64 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha512 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=64 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 32 /dev/urandom 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=a522ecabdf348c0cc4278b01cd57780bfcc9763b7ad27ea3b1b0ddd8494b0037 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha512.XXX 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha512.Gk3 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key a522ecabdf348c0cc4278b01cd57780bfcc9763b7ad27ea3b1b0ddd8494b0037 3 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 a522ecabdf348c0cc4278b01cd57780bfcc9763b7ad27ea3b1b0ddd8494b0037 3 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=a522ecabdf348c0cc4278b01cd57780bfcc9763b7ad27ea3b1b0ddd8494b0037 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=3 00:20:08.634 08:55:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha512.Gk3 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha512.Gk3 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@94 -- # ckeys[0]=/tmp/spdk.key-sha512.Gk3 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # gen_dhchap_key sha256 32 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha256 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=32 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=7059ad3b32ce0c1370b8d01afef96751 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha256.XXX 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha256.XII 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key 7059ad3b32ce0c1370b8d01afef96751 1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 7059ad3b32ce0c1370b8d01afef96751 1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=7059ad3b32ce0c1370b8d01afef96751 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha256.XII 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha256.XII 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # keys[1]=/tmp/spdk.key-sha256.XII 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # gen_dhchap_key sha384 48 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha384 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=48 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=524225691d8ac9f2e82febd386992cadc32de8734230d173 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha384.XXX 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha384.Gf1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key 524225691d8ac9f2e82febd386992cadc32de8734230d173 2 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 524225691d8ac9f2e82febd386992cadc32de8734230d173 2 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=524225691d8ac9f2e82febd386992cadc32de8734230d173 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=2 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha384.Gf1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha384.Gf1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@95 -- # ckeys[1]=/tmp/spdk.key-sha384.Gf1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # gen_dhchap_key sha384 48 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha384 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=48 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=131ffc1dc7731dec62da0982b219400dbd664a4ba46e72c5 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha384.XXX 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha384.OTN 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key 131ffc1dc7731dec62da0982b219400dbd664a4ba46e72c5 2 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 131ffc1dc7731dec62da0982b219400dbd664a4ba46e72c5 2 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=131ffc1dc7731dec62da0982b219400dbd664a4ba46e72c5 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=2 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha384.OTN 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha384.OTN 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # keys[2]=/tmp/spdk.key-sha384.OTN 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # gen_dhchap_key sha256 32 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha256 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=32 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=bc99fe3988bb0ffc33e486c3a64dde57 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha256.XXX 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha256.jmj 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key bc99fe3988bb0ffc33e486c3a64dde57 1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 bc99fe3988bb0ffc33e486c3a64dde57 1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:20:08.634 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=bc99fe3988bb0ffc33e486c3a64dde57 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=1 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha256.jmj 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha256.jmj 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@96 -- # ckeys[2]=/tmp/spdk.key-sha256.jmj 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # gen_dhchap_key sha512 64 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@749 -- # local digest len file key 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@750 -- # local -A digests 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # digest=sha512 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@752 -- # len=64 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # xxd -p -c0 -l 32 /dev/urandom 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@753 -- # key=261e2e769b2e736c894b0ac3fea18d149d64c3e66cfa56bd3f220bb1dcc404ec 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha512.XXX 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha512.AyL 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@755 -- # format_dhchap_key 261e2e769b2e736c894b0ac3fea18d149d64c3e66cfa56bd3f220bb1dcc404ec 3 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@745 -- # format_key DHHC-1 261e2e769b2e736c894b0ac3fea18d149d64c3e66cfa56bd3f220bb1dcc404ec 3 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@728 -- # local prefix key digest 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # key=261e2e769b2e736c894b0ac3fea18d149d64c3e66cfa56bd3f220bb1dcc404ec 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@730 -- # digest=3 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@731 -- # python - 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha512.AyL 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha512.AyL 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # keys[3]=/tmp/spdk.key-sha512.AyL 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@97 -- # ckeys[3]= 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@99 -- # waitforlisten 1866650 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@831 -- # '[' -z 1866650 ']' 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:08.635 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # return 0 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@100 -- # waitforlisten 1866679 /var/tmp/host.sock 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@831 -- # '[' -z 1866679 ']' 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/host.sock 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock...' 00:20:08.635 Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock... 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # return 0 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@101 -- # rpc_cmd 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.2c1 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key0 /tmp/spdk.key-null.2c1 00:20:08.635 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key0 /tmp/spdk.key-null.2c1 00:20:08.948 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha512.Gk3 ]] 00:20:08.948 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.Gk3 00:20:08.948 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:08.948 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:08.948 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:08.948 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey0 /tmp/spdk.key-sha512.Gk3 00:20:08.949 08:55:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey0 /tmp/spdk.key-sha512.Gk3 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-sha256.XII 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key1 /tmp/spdk.key-sha256.XII 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key1 /tmp/spdk.key-sha256.XII 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha384.Gf1 ]] 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.Gf1 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey1 /tmp/spdk.key-sha384.Gf1 00:20:09.234 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey1 /tmp/spdk.key-sha384.Gf1 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha384.OTN 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key2 /tmp/spdk.key-sha384.OTN 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key2 /tmp/spdk.key-sha384.OTN 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n /tmp/spdk.key-sha256.jmj ]] 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@112 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.jmj 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@113 -- # hostrpc keyring_file_add_key ckey2 /tmp/spdk.key-sha256.jmj 00:20:09.495 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey2 /tmp/spdk.key-sha256.jmj 00:20:09.756 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@108 -- # for i in "${!keys[@]}" 00:20:09.756 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@109 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha512.AyL 00:20:09.756 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:09.756 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:09.756 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:09.756 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@110 -- # hostrpc keyring_file_add_key key3 /tmp/spdk.key-sha512.AyL 00:20:09.756 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key3 /tmp/spdk.key-sha512.AyL 00:20:10.017 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@111 -- # [[ -n '' ]] 00:20:10.017 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:20:10.017 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:20:10.017 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:10.017 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:20:10.017 08:55:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 0 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:10.017 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:10.278 00:20:10.278 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:10.278 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:10.278 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:10.539 { 00:20:10.539 "cntlid": 1, 00:20:10.539 "qid": 0, 00:20:10.539 "state": "enabled", 00:20:10.539 "thread": "nvmf_tgt_poll_group_000", 00:20:10.539 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:10.539 "listen_address": { 00:20:10.539 "trtype": "TCP", 00:20:10.539 "adrfam": "IPv4", 00:20:10.539 "traddr": "10.0.0.2", 00:20:10.539 "trsvcid": "4420" 00:20:10.539 }, 00:20:10.539 "peer_address": { 00:20:10.539 "trtype": "TCP", 00:20:10.539 "adrfam": "IPv4", 00:20:10.539 "traddr": "10.0.0.1", 00:20:10.539 "trsvcid": "44328" 00:20:10.539 }, 00:20:10.539 "auth": { 00:20:10.539 "state": "completed", 00:20:10.539 "digest": "sha256", 00:20:10.539 "dhgroup": "null" 00:20:10.539 } 00:20:10.539 } 00:20:10.539 ]' 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:20:10.539 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:10.800 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:10.800 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:10.800 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:10.800 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:10.800 08:56:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:11.740 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 1 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:11.740 08:56:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:12.002 00:20:12.002 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:12.002 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:12.002 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:12.263 { 00:20:12.263 "cntlid": 3, 00:20:12.263 "qid": 0, 00:20:12.263 "state": "enabled", 00:20:12.263 "thread": "nvmf_tgt_poll_group_000", 00:20:12.263 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:12.263 "listen_address": { 00:20:12.263 "trtype": "TCP", 00:20:12.263 "adrfam": "IPv4", 00:20:12.263 "traddr": "10.0.0.2", 00:20:12.263 "trsvcid": "4420" 00:20:12.263 }, 00:20:12.263 "peer_address": { 00:20:12.263 "trtype": "TCP", 00:20:12.263 "adrfam": "IPv4", 00:20:12.263 "traddr": "10.0.0.1", 00:20:12.263 "trsvcid": "44360" 00:20:12.263 }, 00:20:12.263 "auth": { 00:20:12.263 "state": "completed", 00:20:12.263 "digest": "sha256", 00:20:12.263 "dhgroup": "null" 00:20:12.263 } 00:20:12.263 } 00:20:12.263 ]' 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:20:12.263 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:12.524 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:12.524 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:12.524 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:12.524 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:12.524 08:56:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:13.476 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:13.476 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 2 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:13.477 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:13.738 00:20:13.738 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:13.738 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:13.738 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:13.999 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:13.999 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:13.999 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:13.999 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:13.999 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:13.999 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:13.999 { 00:20:13.999 "cntlid": 5, 00:20:13.999 "qid": 0, 00:20:13.999 "state": "enabled", 00:20:13.999 "thread": "nvmf_tgt_poll_group_000", 00:20:13.999 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:13.999 "listen_address": { 00:20:13.999 "trtype": "TCP", 00:20:13.999 "adrfam": "IPv4", 00:20:13.999 "traddr": "10.0.0.2", 00:20:13.999 "trsvcid": "4420" 00:20:13.999 }, 00:20:13.999 "peer_address": { 00:20:13.999 "trtype": "TCP", 00:20:13.999 "adrfam": "IPv4", 00:20:13.999 "traddr": "10.0.0.1", 00:20:13.999 "trsvcid": "44390" 00:20:13.999 }, 00:20:13.999 "auth": { 00:20:13.999 "state": "completed", 00:20:13.999 "digest": "sha256", 00:20:13.999 "dhgroup": "null" 00:20:13.999 } 00:20:13.999 } 00:20:13.999 ]' 00:20:13.999 08:56:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:13.999 08:56:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:13.999 08:56:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:13.999 08:56:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:20:13.999 08:56:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:13.999 08:56:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:13.999 08:56:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:13.999 08:56:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:14.260 08:56:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:14.260 08:56:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:15.202 08:56:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:15.202 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 null 3 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:15.202 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:15.463 00:20:15.463 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:15.463 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:15.463 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:15.725 { 00:20:15.725 "cntlid": 7, 00:20:15.725 "qid": 0, 00:20:15.725 "state": "enabled", 00:20:15.725 "thread": "nvmf_tgt_poll_group_000", 00:20:15.725 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:15.725 "listen_address": { 00:20:15.725 "trtype": "TCP", 00:20:15.725 "adrfam": "IPv4", 00:20:15.725 "traddr": "10.0.0.2", 00:20:15.725 "trsvcid": "4420" 00:20:15.725 }, 00:20:15.725 "peer_address": { 00:20:15.725 "trtype": "TCP", 00:20:15.725 "adrfam": "IPv4", 00:20:15.725 "traddr": "10.0.0.1", 00:20:15.725 "trsvcid": "44412" 00:20:15.725 }, 00:20:15.725 "auth": { 00:20:15.725 "state": "completed", 00:20:15.725 "digest": "sha256", 00:20:15.725 "dhgroup": "null" 00:20:15.725 } 00:20:15.725 } 00:20:15.725 ]' 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:15.725 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:15.986 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:15.986 08:56:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:16.928 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 0 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:16.928 08:56:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:17.189 00:20:17.189 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:17.189 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:17.189 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:17.451 { 00:20:17.451 "cntlid": 9, 00:20:17.451 "qid": 0, 00:20:17.451 "state": "enabled", 00:20:17.451 "thread": "nvmf_tgt_poll_group_000", 00:20:17.451 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:17.451 "listen_address": { 00:20:17.451 "trtype": "TCP", 00:20:17.451 "adrfam": "IPv4", 00:20:17.451 "traddr": "10.0.0.2", 00:20:17.451 "trsvcid": "4420" 00:20:17.451 }, 00:20:17.451 "peer_address": { 00:20:17.451 "trtype": "TCP", 00:20:17.451 "adrfam": "IPv4", 00:20:17.451 "traddr": "10.0.0.1", 00:20:17.451 "trsvcid": "44440" 00:20:17.451 }, 00:20:17.451 "auth": { 00:20:17.451 "state": "completed", 00:20:17.451 "digest": "sha256", 00:20:17.451 "dhgroup": "ffdhe2048" 00:20:17.451 } 00:20:17.451 } 00:20:17.451 ]' 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:17.451 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:17.711 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:17.711 08:56:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:18.282 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:18.282 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:18.282 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:18.282 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:18.282 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:18.282 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:18.282 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:18.282 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:20:18.282 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 1 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:18.543 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:18.803 00:20:18.803 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:18.803 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:18.803 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:19.064 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:19.064 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:19.064 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:19.064 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:19.064 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:19.064 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:19.064 { 00:20:19.064 "cntlid": 11, 00:20:19.064 "qid": 0, 00:20:19.064 "state": "enabled", 00:20:19.064 "thread": "nvmf_tgt_poll_group_000", 00:20:19.064 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:19.064 "listen_address": { 00:20:19.064 "trtype": "TCP", 00:20:19.064 "adrfam": "IPv4", 00:20:19.064 "traddr": "10.0.0.2", 00:20:19.064 "trsvcid": "4420" 00:20:19.064 }, 00:20:19.064 "peer_address": { 00:20:19.064 "trtype": "TCP", 00:20:19.064 "adrfam": "IPv4", 00:20:19.064 "traddr": "10.0.0.1", 00:20:19.064 "trsvcid": "52038" 00:20:19.064 }, 00:20:19.064 "auth": { 00:20:19.064 "state": "completed", 00:20:19.064 "digest": "sha256", 00:20:19.064 "dhgroup": "ffdhe2048" 00:20:19.064 } 00:20:19.064 } 00:20:19.064 ]' 00:20:19.064 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:19.064 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:19.064 08:56:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:19.064 08:56:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:20:19.064 08:56:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:19.064 08:56:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:19.064 08:56:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:19.064 08:56:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:19.326 08:56:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:19.326 08:56:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:20.268 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 2 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:20.268 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:20.529 00:20:20.529 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:20.529 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:20.529 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:20.789 { 00:20:20.789 "cntlid": 13, 00:20:20.789 "qid": 0, 00:20:20.789 "state": "enabled", 00:20:20.789 "thread": "nvmf_tgt_poll_group_000", 00:20:20.789 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:20.789 "listen_address": { 00:20:20.789 "trtype": "TCP", 00:20:20.789 "adrfam": "IPv4", 00:20:20.789 "traddr": "10.0.0.2", 00:20:20.789 "trsvcid": "4420" 00:20:20.789 }, 00:20:20.789 "peer_address": { 00:20:20.789 "trtype": "TCP", 00:20:20.789 "adrfam": "IPv4", 00:20:20.789 "traddr": "10.0.0.1", 00:20:20.789 "trsvcid": "52060" 00:20:20.789 }, 00:20:20.789 "auth": { 00:20:20.789 "state": "completed", 00:20:20.789 "digest": "sha256", 00:20:20.789 "dhgroup": "ffdhe2048" 00:20:20.789 } 00:20:20.789 } 00:20:20.789 ]' 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:20.789 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:20.790 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:20.790 08:56:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:21.050 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:21.050 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:21.991 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe2048 3 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:21.991 08:56:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:22.251 00:20:22.251 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:22.251 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:22.251 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:22.512 { 00:20:22.512 "cntlid": 15, 00:20:22.512 "qid": 0, 00:20:22.512 "state": "enabled", 00:20:22.512 "thread": "nvmf_tgt_poll_group_000", 00:20:22.512 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:22.512 "listen_address": { 00:20:22.512 "trtype": "TCP", 00:20:22.512 "adrfam": "IPv4", 00:20:22.512 "traddr": "10.0.0.2", 00:20:22.512 "trsvcid": "4420" 00:20:22.512 }, 00:20:22.512 "peer_address": { 00:20:22.512 "trtype": "TCP", 00:20:22.512 "adrfam": "IPv4", 00:20:22.512 "traddr": "10.0.0.1", 00:20:22.512 "trsvcid": "52092" 00:20:22.512 }, 00:20:22.512 "auth": { 00:20:22.512 "state": "completed", 00:20:22.512 "digest": "sha256", 00:20:22.512 "dhgroup": "ffdhe2048" 00:20:22.512 } 00:20:22.512 } 00:20:22.512 ]' 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:22.512 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:22.772 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:22.772 08:56:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:23.713 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 0 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:23.713 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:23.714 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:23.714 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:23.974 00:20:23.974 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:23.974 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:23.974 08:56:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:24.235 { 00:20:24.235 "cntlid": 17, 00:20:24.235 "qid": 0, 00:20:24.235 "state": "enabled", 00:20:24.235 "thread": "nvmf_tgt_poll_group_000", 00:20:24.235 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:24.235 "listen_address": { 00:20:24.235 "trtype": "TCP", 00:20:24.235 "adrfam": "IPv4", 00:20:24.235 "traddr": "10.0.0.2", 00:20:24.235 "trsvcid": "4420" 00:20:24.235 }, 00:20:24.235 "peer_address": { 00:20:24.235 "trtype": "TCP", 00:20:24.235 "adrfam": "IPv4", 00:20:24.235 "traddr": "10.0.0.1", 00:20:24.235 "trsvcid": "52122" 00:20:24.235 }, 00:20:24.235 "auth": { 00:20:24.235 "state": "completed", 00:20:24.235 "digest": "sha256", 00:20:24.235 "dhgroup": "ffdhe3072" 00:20:24.235 } 00:20:24.235 } 00:20:24.235 ]' 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:24.235 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:24.495 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:24.495 08:56:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:25.436 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 1 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:25.436 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:25.696 00:20:25.696 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:25.696 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:25.696 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:25.957 { 00:20:25.957 "cntlid": 19, 00:20:25.957 "qid": 0, 00:20:25.957 "state": "enabled", 00:20:25.957 "thread": "nvmf_tgt_poll_group_000", 00:20:25.957 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:25.957 "listen_address": { 00:20:25.957 "trtype": "TCP", 00:20:25.957 "adrfam": "IPv4", 00:20:25.957 "traddr": "10.0.0.2", 00:20:25.957 "trsvcid": "4420" 00:20:25.957 }, 00:20:25.957 "peer_address": { 00:20:25.957 "trtype": "TCP", 00:20:25.957 "adrfam": "IPv4", 00:20:25.957 "traddr": "10.0.0.1", 00:20:25.957 "trsvcid": "52150" 00:20:25.957 }, 00:20:25.957 "auth": { 00:20:25.957 "state": "completed", 00:20:25.957 "digest": "sha256", 00:20:25.957 "dhgroup": "ffdhe3072" 00:20:25.957 } 00:20:25.957 } 00:20:25.957 ]' 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:25.957 08:56:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:26.217 08:56:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:26.217 08:56:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:26.788 08:56:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:26.788 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:26.788 08:56:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:26.788 08:56:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:26.788 08:56:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:26.788 08:56:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:26.788 08:56:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:26.788 08:56:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:20:26.788 08:56:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 2 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:27.050 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:27.310 00:20:27.310 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:27.310 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:27.310 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:27.572 { 00:20:27.572 "cntlid": 21, 00:20:27.572 "qid": 0, 00:20:27.572 "state": "enabled", 00:20:27.572 "thread": "nvmf_tgt_poll_group_000", 00:20:27.572 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:27.572 "listen_address": { 00:20:27.572 "trtype": "TCP", 00:20:27.572 "adrfam": "IPv4", 00:20:27.572 "traddr": "10.0.0.2", 00:20:27.572 "trsvcid": "4420" 00:20:27.572 }, 00:20:27.572 "peer_address": { 00:20:27.572 "trtype": "TCP", 00:20:27.572 "adrfam": "IPv4", 00:20:27.572 "traddr": "10.0.0.1", 00:20:27.572 "trsvcid": "52172" 00:20:27.572 }, 00:20:27.572 "auth": { 00:20:27.572 "state": "completed", 00:20:27.572 "digest": "sha256", 00:20:27.572 "dhgroup": "ffdhe3072" 00:20:27.572 } 00:20:27.572 } 00:20:27.572 ]' 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:27.572 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:27.834 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:27.834 08:56:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:28.776 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe3072 3 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:28.776 08:56:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:29.037 00:20:29.037 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:29.037 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:29.037 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:29.297 { 00:20:29.297 "cntlid": 23, 00:20:29.297 "qid": 0, 00:20:29.297 "state": "enabled", 00:20:29.297 "thread": "nvmf_tgt_poll_group_000", 00:20:29.297 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:29.297 "listen_address": { 00:20:29.297 "trtype": "TCP", 00:20:29.297 "adrfam": "IPv4", 00:20:29.297 "traddr": "10.0.0.2", 00:20:29.297 "trsvcid": "4420" 00:20:29.297 }, 00:20:29.297 "peer_address": { 00:20:29.297 "trtype": "TCP", 00:20:29.297 "adrfam": "IPv4", 00:20:29.297 "traddr": "10.0.0.1", 00:20:29.297 "trsvcid": "58840" 00:20:29.297 }, 00:20:29.297 "auth": { 00:20:29.297 "state": "completed", 00:20:29.297 "digest": "sha256", 00:20:29.297 "dhgroup": "ffdhe3072" 00:20:29.297 } 00:20:29.297 } 00:20:29.297 ]' 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:29.297 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:29.558 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:29.558 08:56:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:30.129 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:30.390 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 0 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:30.390 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:30.391 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:30.652 00:20:30.652 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:30.652 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:30.652 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:30.914 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:30.914 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:30.914 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:30.914 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:30.914 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:30.914 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:30.914 { 00:20:30.914 "cntlid": 25, 00:20:30.914 "qid": 0, 00:20:30.914 "state": "enabled", 00:20:30.914 "thread": "nvmf_tgt_poll_group_000", 00:20:30.914 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:30.914 "listen_address": { 00:20:30.914 "trtype": "TCP", 00:20:30.914 "adrfam": "IPv4", 00:20:30.914 "traddr": "10.0.0.2", 00:20:30.914 "trsvcid": "4420" 00:20:30.914 }, 00:20:30.914 "peer_address": { 00:20:30.914 "trtype": "TCP", 00:20:30.914 "adrfam": "IPv4", 00:20:30.914 "traddr": "10.0.0.1", 00:20:30.914 "trsvcid": "58876" 00:20:30.914 }, 00:20:30.914 "auth": { 00:20:30.914 "state": "completed", 00:20:30.914 "digest": "sha256", 00:20:30.914 "dhgroup": "ffdhe4096" 00:20:30.914 } 00:20:30.914 } 00:20:30.914 ]' 00:20:30.914 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:30.914 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:30.914 08:56:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:30.914 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:20:30.914 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:31.176 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:31.176 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:31.176 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:31.176 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:31.176 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:32.120 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:32.120 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:32.120 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:32.120 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:32.120 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:32.120 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:32.120 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:32.120 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:20:32.120 08:56:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 1 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:32.120 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:32.381 00:20:32.381 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:32.381 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:32.381 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:32.642 { 00:20:32.642 "cntlid": 27, 00:20:32.642 "qid": 0, 00:20:32.642 "state": "enabled", 00:20:32.642 "thread": "nvmf_tgt_poll_group_000", 00:20:32.642 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:32.642 "listen_address": { 00:20:32.642 "trtype": "TCP", 00:20:32.642 "adrfam": "IPv4", 00:20:32.642 "traddr": "10.0.0.2", 00:20:32.642 "trsvcid": "4420" 00:20:32.642 }, 00:20:32.642 "peer_address": { 00:20:32.642 "trtype": "TCP", 00:20:32.642 "adrfam": "IPv4", 00:20:32.642 "traddr": "10.0.0.1", 00:20:32.642 "trsvcid": "58892" 00:20:32.642 }, 00:20:32.642 "auth": { 00:20:32.642 "state": "completed", 00:20:32.642 "digest": "sha256", 00:20:32.642 "dhgroup": "ffdhe4096" 00:20:32.642 } 00:20:32.642 } 00:20:32.642 ]' 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:32.642 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:32.904 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:32.904 08:56:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:33.847 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 2 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:33.847 08:56:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:34.107 00:20:34.107 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:34.107 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:34.107 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:34.368 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:34.369 { 00:20:34.369 "cntlid": 29, 00:20:34.369 "qid": 0, 00:20:34.369 "state": "enabled", 00:20:34.369 "thread": "nvmf_tgt_poll_group_000", 00:20:34.369 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:34.369 "listen_address": { 00:20:34.369 "trtype": "TCP", 00:20:34.369 "adrfam": "IPv4", 00:20:34.369 "traddr": "10.0.0.2", 00:20:34.369 "trsvcid": "4420" 00:20:34.369 }, 00:20:34.369 "peer_address": { 00:20:34.369 "trtype": "TCP", 00:20:34.369 "adrfam": "IPv4", 00:20:34.369 "traddr": "10.0.0.1", 00:20:34.369 "trsvcid": "58918" 00:20:34.369 }, 00:20:34.369 "auth": { 00:20:34.369 "state": "completed", 00:20:34.369 "digest": "sha256", 00:20:34.369 "dhgroup": "ffdhe4096" 00:20:34.369 } 00:20:34.369 } 00:20:34.369 ]' 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:34.369 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:34.630 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:34.630 08:56:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:35.203 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:35.203 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:35.203 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:35.203 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:35.203 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:35.203 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:35.203 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:35.203 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:20:35.203 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe4096 3 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:35.464 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:35.724 00:20:35.725 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:35.725 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:35.725 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:35.985 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:35.985 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:35.985 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:35.985 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:35.985 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:35.985 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:35.985 { 00:20:35.985 "cntlid": 31, 00:20:35.985 "qid": 0, 00:20:35.985 "state": "enabled", 00:20:35.985 "thread": "nvmf_tgt_poll_group_000", 00:20:35.985 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:35.985 "listen_address": { 00:20:35.985 "trtype": "TCP", 00:20:35.985 "adrfam": "IPv4", 00:20:35.985 "traddr": "10.0.0.2", 00:20:35.985 "trsvcid": "4420" 00:20:35.985 }, 00:20:35.985 "peer_address": { 00:20:35.985 "trtype": "TCP", 00:20:35.985 "adrfam": "IPv4", 00:20:35.985 "traddr": "10.0.0.1", 00:20:35.985 "trsvcid": "58942" 00:20:35.985 }, 00:20:35.985 "auth": { 00:20:35.985 "state": "completed", 00:20:35.985 "digest": "sha256", 00:20:35.985 "dhgroup": "ffdhe4096" 00:20:35.985 } 00:20:35.985 } 00:20:35.985 ]' 00:20:35.985 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:35.985 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:35.985 08:56:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:35.985 08:56:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:20:35.985 08:56:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:35.985 08:56:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:35.985 08:56:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:35.985 08:56:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:36.246 08:56:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:36.246 08:56:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:37.189 08:56:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:37.189 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 0 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:37.189 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:37.449 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:37.710 { 00:20:37.710 "cntlid": 33, 00:20:37.710 "qid": 0, 00:20:37.710 "state": "enabled", 00:20:37.710 "thread": "nvmf_tgt_poll_group_000", 00:20:37.710 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:37.710 "listen_address": { 00:20:37.710 "trtype": "TCP", 00:20:37.710 "adrfam": "IPv4", 00:20:37.710 "traddr": "10.0.0.2", 00:20:37.710 "trsvcid": "4420" 00:20:37.710 }, 00:20:37.710 "peer_address": { 00:20:37.710 "trtype": "TCP", 00:20:37.710 "adrfam": "IPv4", 00:20:37.710 "traddr": "10.0.0.1", 00:20:37.710 "trsvcid": "58972" 00:20:37.710 }, 00:20:37.710 "auth": { 00:20:37.710 "state": "completed", 00:20:37.710 "digest": "sha256", 00:20:37.710 "dhgroup": "ffdhe6144" 00:20:37.710 } 00:20:37.710 } 00:20:37.710 ]' 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:37.710 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:37.971 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:20:37.971 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:37.971 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:37.971 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:37.971 08:56:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:37.971 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:37.971 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:38.913 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 1 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:38.913 08:56:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:39.484 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:39.484 { 00:20:39.484 "cntlid": 35, 00:20:39.484 "qid": 0, 00:20:39.484 "state": "enabled", 00:20:39.484 "thread": "nvmf_tgt_poll_group_000", 00:20:39.484 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:39.484 "listen_address": { 00:20:39.484 "trtype": "TCP", 00:20:39.484 "adrfam": "IPv4", 00:20:39.484 "traddr": "10.0.0.2", 00:20:39.484 "trsvcid": "4420" 00:20:39.484 }, 00:20:39.484 "peer_address": { 00:20:39.484 "trtype": "TCP", 00:20:39.484 "adrfam": "IPv4", 00:20:39.484 "traddr": "10.0.0.1", 00:20:39.484 "trsvcid": "41590" 00:20:39.484 }, 00:20:39.484 "auth": { 00:20:39.484 "state": "completed", 00:20:39.484 "digest": "sha256", 00:20:39.484 "dhgroup": "ffdhe6144" 00:20:39.484 } 00:20:39.484 } 00:20:39.484 ]' 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:39.484 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:39.747 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:20:39.747 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:39.747 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:39.747 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:39.747 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:39.747 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:39.747 08:56:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:40.690 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 2 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:40.690 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:40.951 08:56:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:41.212 00:20:41.212 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:41.212 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:41.212 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:41.473 { 00:20:41.473 "cntlid": 37, 00:20:41.473 "qid": 0, 00:20:41.473 "state": "enabled", 00:20:41.473 "thread": "nvmf_tgt_poll_group_000", 00:20:41.473 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:41.473 "listen_address": { 00:20:41.473 "trtype": "TCP", 00:20:41.473 "adrfam": "IPv4", 00:20:41.473 "traddr": "10.0.0.2", 00:20:41.473 "trsvcid": "4420" 00:20:41.473 }, 00:20:41.473 "peer_address": { 00:20:41.473 "trtype": "TCP", 00:20:41.473 "adrfam": "IPv4", 00:20:41.473 "traddr": "10.0.0.1", 00:20:41.473 "trsvcid": "41626" 00:20:41.473 }, 00:20:41.473 "auth": { 00:20:41.473 "state": "completed", 00:20:41.473 "digest": "sha256", 00:20:41.473 "dhgroup": "ffdhe6144" 00:20:41.473 } 00:20:41.473 } 00:20:41.473 ]' 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:41.473 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:41.735 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:41.735 08:56:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:42.307 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:42.307 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:42.307 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:42.307 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:42.307 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:42.307 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:42.307 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:42.307 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:20:42.307 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe6144 3 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:42.568 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:43.143 00:20:43.143 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:43.143 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:43.143 08:56:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:43.143 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:43.143 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:43.143 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:43.143 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:43.143 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:43.143 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:43.143 { 00:20:43.143 "cntlid": 39, 00:20:43.143 "qid": 0, 00:20:43.143 "state": "enabled", 00:20:43.143 "thread": "nvmf_tgt_poll_group_000", 00:20:43.143 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:43.143 "listen_address": { 00:20:43.143 "trtype": "TCP", 00:20:43.143 "adrfam": "IPv4", 00:20:43.143 "traddr": "10.0.0.2", 00:20:43.143 "trsvcid": "4420" 00:20:43.143 }, 00:20:43.143 "peer_address": { 00:20:43.143 "trtype": "TCP", 00:20:43.143 "adrfam": "IPv4", 00:20:43.143 "traddr": "10.0.0.1", 00:20:43.143 "trsvcid": "41672" 00:20:43.143 }, 00:20:43.143 "auth": { 00:20:43.143 "state": "completed", 00:20:43.143 "digest": "sha256", 00:20:43.143 "dhgroup": "ffdhe6144" 00:20:43.143 } 00:20:43.143 } 00:20:43.143 ]' 00:20:43.143 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:43.143 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:43.143 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:43.405 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:20:43.405 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:43.405 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:43.405 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:43.405 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:43.405 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:43.405 08:56:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:44.348 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 0 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:44.348 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:44.921 00:20:44.921 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:44.921 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:44.921 08:56:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:45.182 { 00:20:45.182 "cntlid": 41, 00:20:45.182 "qid": 0, 00:20:45.182 "state": "enabled", 00:20:45.182 "thread": "nvmf_tgt_poll_group_000", 00:20:45.182 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:45.182 "listen_address": { 00:20:45.182 "trtype": "TCP", 00:20:45.182 "adrfam": "IPv4", 00:20:45.182 "traddr": "10.0.0.2", 00:20:45.182 "trsvcid": "4420" 00:20:45.182 }, 00:20:45.182 "peer_address": { 00:20:45.182 "trtype": "TCP", 00:20:45.182 "adrfam": "IPv4", 00:20:45.182 "traddr": "10.0.0.1", 00:20:45.182 "trsvcid": "41696" 00:20:45.182 }, 00:20:45.182 "auth": { 00:20:45.182 "state": "completed", 00:20:45.182 "digest": "sha256", 00:20:45.182 "dhgroup": "ffdhe8192" 00:20:45.182 } 00:20:45.182 } 00:20:45.182 ]' 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:45.182 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:45.443 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:45.443 08:56:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:46.013 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:46.013 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:46.013 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:46.013 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:46.013 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:46.013 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:46.013 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:46.013 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:20:46.013 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:20:46.274 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 1 00:20:46.274 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:46.274 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:46.275 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:20:46.275 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:20:46.275 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:46.275 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:46.275 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:46.275 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:46.275 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:46.275 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:46.275 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:46.275 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:46.845 00:20:46.845 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:46.845 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:46.845 08:56:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:47.105 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:47.105 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:47.105 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:47.105 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:47.105 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:47.105 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:47.105 { 00:20:47.105 "cntlid": 43, 00:20:47.105 "qid": 0, 00:20:47.105 "state": "enabled", 00:20:47.105 "thread": "nvmf_tgt_poll_group_000", 00:20:47.105 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:47.105 "listen_address": { 00:20:47.105 "trtype": "TCP", 00:20:47.105 "adrfam": "IPv4", 00:20:47.105 "traddr": "10.0.0.2", 00:20:47.105 "trsvcid": "4420" 00:20:47.105 }, 00:20:47.105 "peer_address": { 00:20:47.105 "trtype": "TCP", 00:20:47.105 "adrfam": "IPv4", 00:20:47.105 "traddr": "10.0.0.1", 00:20:47.105 "trsvcid": "41716" 00:20:47.105 }, 00:20:47.105 "auth": { 00:20:47.105 "state": "completed", 00:20:47.106 "digest": "sha256", 00:20:47.106 "dhgroup": "ffdhe8192" 00:20:47.106 } 00:20:47.106 } 00:20:47.106 ]' 00:20:47.106 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:47.106 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:47.106 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:47.106 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:20:47.106 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:47.106 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:47.106 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:47.106 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:47.391 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:47.391 08:56:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:48.064 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:48.064 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:48.064 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:48.064 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:48.064 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:48.064 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:48.064 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:48.064 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:20:48.064 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 2 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:48.324 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:48.895 00:20:48.895 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:48.895 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:48.895 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:48.895 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:48.895 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:48.895 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:48.895 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:48.895 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:48.895 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:48.895 { 00:20:48.895 "cntlid": 45, 00:20:48.895 "qid": 0, 00:20:48.895 "state": "enabled", 00:20:48.895 "thread": "nvmf_tgt_poll_group_000", 00:20:48.895 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:48.895 "listen_address": { 00:20:48.895 "trtype": "TCP", 00:20:48.895 "adrfam": "IPv4", 00:20:48.895 "traddr": "10.0.0.2", 00:20:48.895 "trsvcid": "4420" 00:20:48.895 }, 00:20:48.895 "peer_address": { 00:20:48.895 "trtype": "TCP", 00:20:48.895 "adrfam": "IPv4", 00:20:48.895 "traddr": "10.0.0.1", 00:20:48.895 "trsvcid": "46124" 00:20:48.895 }, 00:20:48.895 "auth": { 00:20:48.895 "state": "completed", 00:20:48.895 "digest": "sha256", 00:20:48.895 "dhgroup": "ffdhe8192" 00:20:48.895 } 00:20:48.895 } 00:20:48.895 ]' 00:20:48.895 08:56:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:49.155 08:56:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:49.155 08:56:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:49.155 08:56:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:20:49.155 08:56:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:49.155 08:56:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:49.155 08:56:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:49.155 08:56:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:49.415 08:56:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:49.415 08:56:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:49.989 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:49.989 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:49.989 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:49.989 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:49.989 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:49.989 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:49.989 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:49.989 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:20:49.989 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha256 ffdhe8192 3 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha256 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:50.249 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:50.820 00:20:50.820 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:50.820 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:50.820 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:51.080 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:51.080 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:51.080 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:51.080 08:56:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:51.080 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:51.080 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:51.080 { 00:20:51.080 "cntlid": 47, 00:20:51.080 "qid": 0, 00:20:51.080 "state": "enabled", 00:20:51.080 "thread": "nvmf_tgt_poll_group_000", 00:20:51.080 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:51.080 "listen_address": { 00:20:51.080 "trtype": "TCP", 00:20:51.080 "adrfam": "IPv4", 00:20:51.080 "traddr": "10.0.0.2", 00:20:51.080 "trsvcid": "4420" 00:20:51.080 }, 00:20:51.080 "peer_address": { 00:20:51.080 "trtype": "TCP", 00:20:51.080 "adrfam": "IPv4", 00:20:51.080 "traddr": "10.0.0.1", 00:20:51.080 "trsvcid": "46140" 00:20:51.080 }, 00:20:51.080 "auth": { 00:20:51.080 "state": "completed", 00:20:51.080 "digest": "sha256", 00:20:51.080 "dhgroup": "ffdhe8192" 00:20:51.080 } 00:20:51.080 } 00:20:51.080 ]' 00:20:51.080 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:51.080 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:20:51.080 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:51.080 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:20:51.080 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:51.080 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:51.080 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:51.080 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:51.341 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:51.341 08:56:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:52.283 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 0 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:52.283 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:52.544 00:20:52.544 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:52.544 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:52.544 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:52.806 { 00:20:52.806 "cntlid": 49, 00:20:52.806 "qid": 0, 00:20:52.806 "state": "enabled", 00:20:52.806 "thread": "nvmf_tgt_poll_group_000", 00:20:52.806 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:52.806 "listen_address": { 00:20:52.806 "trtype": "TCP", 00:20:52.806 "adrfam": "IPv4", 00:20:52.806 "traddr": "10.0.0.2", 00:20:52.806 "trsvcid": "4420" 00:20:52.806 }, 00:20:52.806 "peer_address": { 00:20:52.806 "trtype": "TCP", 00:20:52.806 "adrfam": "IPv4", 00:20:52.806 "traddr": "10.0.0.1", 00:20:52.806 "trsvcid": "46168" 00:20:52.806 }, 00:20:52.806 "auth": { 00:20:52.806 "state": "completed", 00:20:52.806 "digest": "sha384", 00:20:52.806 "dhgroup": "null" 00:20:52.806 } 00:20:52.806 } 00:20:52.806 ]' 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:52.806 08:56:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:53.066 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:53.066 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:54.009 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 1 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:54.009 08:56:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:20:54.270 00:20:54.270 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:54.270 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:54.270 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:54.530 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:54.530 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:54.530 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:54.530 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:54.531 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:54.531 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:54.531 { 00:20:54.531 "cntlid": 51, 00:20:54.531 "qid": 0, 00:20:54.531 "state": "enabled", 00:20:54.531 "thread": "nvmf_tgt_poll_group_000", 00:20:54.531 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:54.531 "listen_address": { 00:20:54.531 "trtype": "TCP", 00:20:54.531 "adrfam": "IPv4", 00:20:54.531 "traddr": "10.0.0.2", 00:20:54.531 "trsvcid": "4420" 00:20:54.531 }, 00:20:54.531 "peer_address": { 00:20:54.531 "trtype": "TCP", 00:20:54.531 "adrfam": "IPv4", 00:20:54.531 "traddr": "10.0.0.1", 00:20:54.531 "trsvcid": "46192" 00:20:54.531 }, 00:20:54.531 "auth": { 00:20:54.531 "state": "completed", 00:20:54.531 "digest": "sha384", 00:20:54.531 "dhgroup": "null" 00:20:54.531 } 00:20:54.531 } 00:20:54.531 ]' 00:20:54.531 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:54.531 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:20:54.531 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:54.531 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:20:54.531 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:54.531 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:54.531 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:54.531 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:54.790 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:54.790 08:56:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:20:55.361 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:55.621 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 2 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:55.621 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:20:55.881 00:20:55.881 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:55.881 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:55.881 08:56:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:56.142 { 00:20:56.142 "cntlid": 53, 00:20:56.142 "qid": 0, 00:20:56.142 "state": "enabled", 00:20:56.142 "thread": "nvmf_tgt_poll_group_000", 00:20:56.142 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:56.142 "listen_address": { 00:20:56.142 "trtype": "TCP", 00:20:56.142 "adrfam": "IPv4", 00:20:56.142 "traddr": "10.0.0.2", 00:20:56.142 "trsvcid": "4420" 00:20:56.142 }, 00:20:56.142 "peer_address": { 00:20:56.142 "trtype": "TCP", 00:20:56.142 "adrfam": "IPv4", 00:20:56.142 "traddr": "10.0.0.1", 00:20:56.142 "trsvcid": "46218" 00:20:56.142 }, 00:20:56.142 "auth": { 00:20:56.142 "state": "completed", 00:20:56.142 "digest": "sha384", 00:20:56.142 "dhgroup": "null" 00:20:56.142 } 00:20:56.142 } 00:20:56.142 ]' 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:56.142 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:56.403 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:56.403 08:56:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:57.344 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 null 3 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:57.344 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:20:57.604 00:20:57.604 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:57.604 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:57.604 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:57.865 { 00:20:57.865 "cntlid": 55, 00:20:57.865 "qid": 0, 00:20:57.865 "state": "enabled", 00:20:57.865 "thread": "nvmf_tgt_poll_group_000", 00:20:57.865 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:57.865 "listen_address": { 00:20:57.865 "trtype": "TCP", 00:20:57.865 "adrfam": "IPv4", 00:20:57.865 "traddr": "10.0.0.2", 00:20:57.865 "trsvcid": "4420" 00:20:57.865 }, 00:20:57.865 "peer_address": { 00:20:57.865 "trtype": "TCP", 00:20:57.865 "adrfam": "IPv4", 00:20:57.865 "traddr": "10.0.0.1", 00:20:57.865 "trsvcid": "46248" 00:20:57.865 }, 00:20:57.865 "auth": { 00:20:57.865 "state": "completed", 00:20:57.865 "digest": "sha384", 00:20:57.865 "dhgroup": "null" 00:20:57.865 } 00:20:57.865 } 00:20:57.865 ]' 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:57.865 08:56:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:58.125 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:58.125 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:20:59.065 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:20:59.065 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:20:59.066 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:20:59.066 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:59.066 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:59.066 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:59.066 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:20:59.066 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:20:59.066 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:20:59.066 08:56:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 0 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:59.066 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:20:59.326 00:20:59.326 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:20:59.326 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:20:59.326 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:20:59.586 { 00:20:59.586 "cntlid": 57, 00:20:59.586 "qid": 0, 00:20:59.586 "state": "enabled", 00:20:59.586 "thread": "nvmf_tgt_poll_group_000", 00:20:59.586 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:20:59.586 "listen_address": { 00:20:59.586 "trtype": "TCP", 00:20:59.586 "adrfam": "IPv4", 00:20:59.586 "traddr": "10.0.0.2", 00:20:59.586 "trsvcid": "4420" 00:20:59.586 }, 00:20:59.586 "peer_address": { 00:20:59.586 "trtype": "TCP", 00:20:59.586 "adrfam": "IPv4", 00:20:59.586 "traddr": "10.0.0.1", 00:20:59.586 "trsvcid": "37398" 00:20:59.586 }, 00:20:59.586 "auth": { 00:20:59.586 "state": "completed", 00:20:59.586 "digest": "sha384", 00:20:59.586 "dhgroup": "ffdhe2048" 00:20:59.586 } 00:20:59.586 } 00:20:59.586 ]' 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:20:59.586 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:20:59.845 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:20:59.845 08:56:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:00.785 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 1 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:00.785 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:21:00.786 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:21:00.786 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:00.786 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:00.786 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:00.786 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:00.786 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:00.786 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:00.786 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:00.786 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:01.047 00:21:01.047 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:01.047 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:01.047 08:56:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:01.047 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:01.047 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:01.047 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:01.047 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:01.308 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:01.308 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:01.308 { 00:21:01.308 "cntlid": 59, 00:21:01.308 "qid": 0, 00:21:01.308 "state": "enabled", 00:21:01.308 "thread": "nvmf_tgt_poll_group_000", 00:21:01.308 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:01.308 "listen_address": { 00:21:01.308 "trtype": "TCP", 00:21:01.308 "adrfam": "IPv4", 00:21:01.308 "traddr": "10.0.0.2", 00:21:01.308 "trsvcid": "4420" 00:21:01.308 }, 00:21:01.308 "peer_address": { 00:21:01.308 "trtype": "TCP", 00:21:01.308 "adrfam": "IPv4", 00:21:01.308 "traddr": "10.0.0.1", 00:21:01.308 "trsvcid": "37442" 00:21:01.308 }, 00:21:01.308 "auth": { 00:21:01.308 "state": "completed", 00:21:01.308 "digest": "sha384", 00:21:01.308 "dhgroup": "ffdhe2048" 00:21:01.308 } 00:21:01.308 } 00:21:01.308 ]' 00:21:01.308 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:01.308 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:01.308 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:01.308 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:21:01.308 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:01.308 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:01.308 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:01.308 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:01.569 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:01.569 08:56:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:02.140 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:02.401 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 2 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:02.401 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:02.663 00:21:02.663 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:02.663 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:02.663 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:02.923 { 00:21:02.923 "cntlid": 61, 00:21:02.923 "qid": 0, 00:21:02.923 "state": "enabled", 00:21:02.923 "thread": "nvmf_tgt_poll_group_000", 00:21:02.923 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:02.923 "listen_address": { 00:21:02.923 "trtype": "TCP", 00:21:02.923 "adrfam": "IPv4", 00:21:02.923 "traddr": "10.0.0.2", 00:21:02.923 "trsvcid": "4420" 00:21:02.923 }, 00:21:02.923 "peer_address": { 00:21:02.923 "trtype": "TCP", 00:21:02.923 "adrfam": "IPv4", 00:21:02.923 "traddr": "10.0.0.1", 00:21:02.923 "trsvcid": "37474" 00:21:02.923 }, 00:21:02.923 "auth": { 00:21:02.923 "state": "completed", 00:21:02.923 "digest": "sha384", 00:21:02.923 "dhgroup": "ffdhe2048" 00:21:02.923 } 00:21:02.923 } 00:21:02.923 ]' 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:21:02.923 08:56:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:02.923 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:02.923 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:02.923 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:03.184 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:03.184 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:04.127 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:04.127 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:04.127 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:04.127 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.127 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:04.127 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.127 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:04.127 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:21:04.127 08:56:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe2048 3 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:04.127 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:04.388 00:21:04.388 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:04.388 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:04.388 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:04.388 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:04.388 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:04.388 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:04.388 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:04.649 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:04.649 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:04.649 { 00:21:04.649 "cntlid": 63, 00:21:04.649 "qid": 0, 00:21:04.649 "state": "enabled", 00:21:04.649 "thread": "nvmf_tgt_poll_group_000", 00:21:04.649 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:04.649 "listen_address": { 00:21:04.649 "trtype": "TCP", 00:21:04.649 "adrfam": "IPv4", 00:21:04.649 "traddr": "10.0.0.2", 00:21:04.649 "trsvcid": "4420" 00:21:04.649 }, 00:21:04.649 "peer_address": { 00:21:04.649 "trtype": "TCP", 00:21:04.649 "adrfam": "IPv4", 00:21:04.649 "traddr": "10.0.0.1", 00:21:04.649 "trsvcid": "37516" 00:21:04.649 }, 00:21:04.649 "auth": { 00:21:04.649 "state": "completed", 00:21:04.649 "digest": "sha384", 00:21:04.649 "dhgroup": "ffdhe2048" 00:21:04.649 } 00:21:04.649 } 00:21:04.649 ]' 00:21:04.649 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:04.649 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:04.649 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:04.649 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:21:04.649 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:04.649 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:04.649 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:04.649 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:04.909 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:04.909 08:56:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:05.480 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:05.480 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:05.480 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:05.480 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:05.480 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:05.480 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:05.480 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:21:05.480 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:05.480 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:21:05.480 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 0 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:05.741 08:56:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:06.002 00:21:06.002 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:06.002 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:06.002 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:06.263 { 00:21:06.263 "cntlid": 65, 00:21:06.263 "qid": 0, 00:21:06.263 "state": "enabled", 00:21:06.263 "thread": "nvmf_tgt_poll_group_000", 00:21:06.263 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:06.263 "listen_address": { 00:21:06.263 "trtype": "TCP", 00:21:06.263 "adrfam": "IPv4", 00:21:06.263 "traddr": "10.0.0.2", 00:21:06.263 "trsvcid": "4420" 00:21:06.263 }, 00:21:06.263 "peer_address": { 00:21:06.263 "trtype": "TCP", 00:21:06.263 "adrfam": "IPv4", 00:21:06.263 "traddr": "10.0.0.1", 00:21:06.263 "trsvcid": "37540" 00:21:06.263 }, 00:21:06.263 "auth": { 00:21:06.263 "state": "completed", 00:21:06.263 "digest": "sha384", 00:21:06.263 "dhgroup": "ffdhe3072" 00:21:06.263 } 00:21:06.263 } 00:21:06.263 ]' 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:06.263 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:06.523 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:06.523 08:56:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:07.465 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 1 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:07.465 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:07.726 00:21:07.726 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:07.727 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:07.727 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:07.987 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:07.987 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:07.987 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:07.987 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:07.987 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:07.987 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:07.987 { 00:21:07.987 "cntlid": 67, 00:21:07.987 "qid": 0, 00:21:07.987 "state": "enabled", 00:21:07.987 "thread": "nvmf_tgt_poll_group_000", 00:21:07.987 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:07.987 "listen_address": { 00:21:07.987 "trtype": "TCP", 00:21:07.987 "adrfam": "IPv4", 00:21:07.987 "traddr": "10.0.0.2", 00:21:07.987 "trsvcid": "4420" 00:21:07.987 }, 00:21:07.987 "peer_address": { 00:21:07.987 "trtype": "TCP", 00:21:07.987 "adrfam": "IPv4", 00:21:07.987 "traddr": "10.0.0.1", 00:21:07.987 "trsvcid": "37558" 00:21:07.987 }, 00:21:07.987 "auth": { 00:21:07.987 "state": "completed", 00:21:07.987 "digest": "sha384", 00:21:07.987 "dhgroup": "ffdhe3072" 00:21:07.987 } 00:21:07.987 } 00:21:07.987 ]' 00:21:07.987 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:07.987 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:07.987 08:56:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:07.987 08:56:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:21:07.987 08:56:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:07.987 08:56:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:07.987 08:56:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:07.987 08:56:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:08.248 08:56:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:08.248 08:56:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:09.191 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 2 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:09.191 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:09.451 00:21:09.451 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:09.451 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:09.451 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:09.713 { 00:21:09.713 "cntlid": 69, 00:21:09.713 "qid": 0, 00:21:09.713 "state": "enabled", 00:21:09.713 "thread": "nvmf_tgt_poll_group_000", 00:21:09.713 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:09.713 "listen_address": { 00:21:09.713 "trtype": "TCP", 00:21:09.713 "adrfam": "IPv4", 00:21:09.713 "traddr": "10.0.0.2", 00:21:09.713 "trsvcid": "4420" 00:21:09.713 }, 00:21:09.713 "peer_address": { 00:21:09.713 "trtype": "TCP", 00:21:09.713 "adrfam": "IPv4", 00:21:09.713 "traddr": "10.0.0.1", 00:21:09.713 "trsvcid": "45904" 00:21:09.713 }, 00:21:09.713 "auth": { 00:21:09.713 "state": "completed", 00:21:09.713 "digest": "sha384", 00:21:09.713 "dhgroup": "ffdhe3072" 00:21:09.713 } 00:21:09.713 } 00:21:09.713 ]' 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:09.713 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:09.977 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:09.977 08:56:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:10.917 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe3072 3 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:10.917 08:57:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:11.178 00:21:11.178 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:11.178 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:11.178 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:11.439 { 00:21:11.439 "cntlid": 71, 00:21:11.439 "qid": 0, 00:21:11.439 "state": "enabled", 00:21:11.439 "thread": "nvmf_tgt_poll_group_000", 00:21:11.439 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:11.439 "listen_address": { 00:21:11.439 "trtype": "TCP", 00:21:11.439 "adrfam": "IPv4", 00:21:11.439 "traddr": "10.0.0.2", 00:21:11.439 "trsvcid": "4420" 00:21:11.439 }, 00:21:11.439 "peer_address": { 00:21:11.439 "trtype": "TCP", 00:21:11.439 "adrfam": "IPv4", 00:21:11.439 "traddr": "10.0.0.1", 00:21:11.439 "trsvcid": "45926" 00:21:11.439 }, 00:21:11.439 "auth": { 00:21:11.439 "state": "completed", 00:21:11.439 "digest": "sha384", 00:21:11.439 "dhgroup": "ffdhe3072" 00:21:11.439 } 00:21:11.439 } 00:21:11.439 ]' 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:11.439 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:11.699 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:11.699 08:57:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:12.270 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:12.531 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 0 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:12.531 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:12.792 00:21:12.792 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:12.792 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:12.792 08:57:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:13.052 { 00:21:13.052 "cntlid": 73, 00:21:13.052 "qid": 0, 00:21:13.052 "state": "enabled", 00:21:13.052 "thread": "nvmf_tgt_poll_group_000", 00:21:13.052 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:13.052 "listen_address": { 00:21:13.052 "trtype": "TCP", 00:21:13.052 "adrfam": "IPv4", 00:21:13.052 "traddr": "10.0.0.2", 00:21:13.052 "trsvcid": "4420" 00:21:13.052 }, 00:21:13.052 "peer_address": { 00:21:13.052 "trtype": "TCP", 00:21:13.052 "adrfam": "IPv4", 00:21:13.052 "traddr": "10.0.0.1", 00:21:13.052 "trsvcid": "45946" 00:21:13.052 }, 00:21:13.052 "auth": { 00:21:13.052 "state": "completed", 00:21:13.052 "digest": "sha384", 00:21:13.052 "dhgroup": "ffdhe4096" 00:21:13.052 } 00:21:13.052 } 00:21:13.052 ]' 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:21:13.052 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:13.313 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:13.313 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:13.313 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:13.313 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:13.313 08:57:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:14.254 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 1 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:14.254 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:14.514 00:21:14.514 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:14.514 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:14.514 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:14.774 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:14.775 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:14.775 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:14.775 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:14.775 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:14.775 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:14.775 { 00:21:14.775 "cntlid": 75, 00:21:14.775 "qid": 0, 00:21:14.775 "state": "enabled", 00:21:14.775 "thread": "nvmf_tgt_poll_group_000", 00:21:14.775 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:14.775 "listen_address": { 00:21:14.775 "trtype": "TCP", 00:21:14.775 "adrfam": "IPv4", 00:21:14.775 "traddr": "10.0.0.2", 00:21:14.775 "trsvcid": "4420" 00:21:14.775 }, 00:21:14.775 "peer_address": { 00:21:14.775 "trtype": "TCP", 00:21:14.775 "adrfam": "IPv4", 00:21:14.775 "traddr": "10.0.0.1", 00:21:14.775 "trsvcid": "45976" 00:21:14.775 }, 00:21:14.775 "auth": { 00:21:14.775 "state": "completed", 00:21:14.775 "digest": "sha384", 00:21:14.775 "dhgroup": "ffdhe4096" 00:21:14.775 } 00:21:14.775 } 00:21:14.775 ]' 00:21:14.775 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:14.775 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:14.775 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:14.775 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:21:14.775 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:15.036 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:15.036 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:15.036 08:57:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:15.036 08:57:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:15.036 08:57:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:15.977 08:57:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:15.977 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:15.977 08:57:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:15.977 08:57:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:15.977 08:57:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:15.977 08:57:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:15.977 08:57:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:15.977 08:57:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:21:15.977 08:57:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:21:15.977 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 2 00:21:15.977 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:15.977 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:15.977 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:21:15.977 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:21:15.977 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:15.978 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:15.978 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:15.978 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:15.978 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:15.978 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:15.978 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:15.978 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:16.239 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:16.500 { 00:21:16.500 "cntlid": 77, 00:21:16.500 "qid": 0, 00:21:16.500 "state": "enabled", 00:21:16.500 "thread": "nvmf_tgt_poll_group_000", 00:21:16.500 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:16.500 "listen_address": { 00:21:16.500 "trtype": "TCP", 00:21:16.500 "adrfam": "IPv4", 00:21:16.500 "traddr": "10.0.0.2", 00:21:16.500 "trsvcid": "4420" 00:21:16.500 }, 00:21:16.500 "peer_address": { 00:21:16.500 "trtype": "TCP", 00:21:16.500 "adrfam": "IPv4", 00:21:16.500 "traddr": "10.0.0.1", 00:21:16.500 "trsvcid": "46016" 00:21:16.500 }, 00:21:16.500 "auth": { 00:21:16.500 "state": "completed", 00:21:16.500 "digest": "sha384", 00:21:16.500 "dhgroup": "ffdhe4096" 00:21:16.500 } 00:21:16.500 } 00:21:16.500 ]' 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:16.500 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:16.761 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:21:16.761 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:16.761 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:16.761 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:16.761 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:16.761 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:16.761 08:57:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:17.702 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:17.703 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe4096 3 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:17.703 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:17.963 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:17.963 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:21:17.963 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:17.963 08:57:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:18.224 00:21:18.224 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:18.224 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:18.224 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:18.224 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:18.224 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:18.224 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:18.224 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:18.224 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:18.224 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:18.224 { 00:21:18.224 "cntlid": 79, 00:21:18.224 "qid": 0, 00:21:18.224 "state": "enabled", 00:21:18.224 "thread": "nvmf_tgt_poll_group_000", 00:21:18.224 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:18.224 "listen_address": { 00:21:18.224 "trtype": "TCP", 00:21:18.224 "adrfam": "IPv4", 00:21:18.224 "traddr": "10.0.0.2", 00:21:18.224 "trsvcid": "4420" 00:21:18.224 }, 00:21:18.224 "peer_address": { 00:21:18.224 "trtype": "TCP", 00:21:18.224 "adrfam": "IPv4", 00:21:18.224 "traddr": "10.0.0.1", 00:21:18.224 "trsvcid": "46038" 00:21:18.224 }, 00:21:18.224 "auth": { 00:21:18.224 "state": "completed", 00:21:18.224 "digest": "sha384", 00:21:18.224 "dhgroup": "ffdhe4096" 00:21:18.224 } 00:21:18.224 } 00:21:18.224 ]' 00:21:18.224 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:18.485 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:18.485 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:18.485 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:21:18.485 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:18.485 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:18.485 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:18.485 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:18.746 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:18.746 08:57:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:19.319 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:19.319 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:19.319 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:19.319 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:19.319 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:19.319 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:19.319 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:21:19.319 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:19.319 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:21:19.319 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 0 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:19.580 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:19.839 00:21:19.839 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:19.839 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:19.839 08:57:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:20.100 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:20.100 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:20.100 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:20.100 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:20.100 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:20.100 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:20.100 { 00:21:20.100 "cntlid": 81, 00:21:20.100 "qid": 0, 00:21:20.100 "state": "enabled", 00:21:20.100 "thread": "nvmf_tgt_poll_group_000", 00:21:20.100 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:20.100 "listen_address": { 00:21:20.100 "trtype": "TCP", 00:21:20.100 "adrfam": "IPv4", 00:21:20.100 "traddr": "10.0.0.2", 00:21:20.100 "trsvcid": "4420" 00:21:20.100 }, 00:21:20.100 "peer_address": { 00:21:20.100 "trtype": "TCP", 00:21:20.100 "adrfam": "IPv4", 00:21:20.100 "traddr": "10.0.0.1", 00:21:20.100 "trsvcid": "36494" 00:21:20.100 }, 00:21:20.100 "auth": { 00:21:20.100 "state": "completed", 00:21:20.100 "digest": "sha384", 00:21:20.100 "dhgroup": "ffdhe6144" 00:21:20.100 } 00:21:20.100 } 00:21:20.100 ]' 00:21:20.100 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:20.100 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:20.100 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:20.361 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:21:20.361 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:20.361 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:20.361 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:20.361 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:20.361 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:20.361 08:57:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:21.302 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 1 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:21.302 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:21.562 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:21.562 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:21.562 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:21.562 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:21.823 00:21:21.823 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:21.823 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:21.823 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:22.083 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:22.083 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:22.083 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:22.083 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:22.083 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:22.083 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:22.083 { 00:21:22.083 "cntlid": 83, 00:21:22.083 "qid": 0, 00:21:22.083 "state": "enabled", 00:21:22.083 "thread": "nvmf_tgt_poll_group_000", 00:21:22.083 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:22.083 "listen_address": { 00:21:22.083 "trtype": "TCP", 00:21:22.083 "adrfam": "IPv4", 00:21:22.083 "traddr": "10.0.0.2", 00:21:22.083 "trsvcid": "4420" 00:21:22.083 }, 00:21:22.083 "peer_address": { 00:21:22.083 "trtype": "TCP", 00:21:22.083 "adrfam": "IPv4", 00:21:22.083 "traddr": "10.0.0.1", 00:21:22.083 "trsvcid": "36514" 00:21:22.083 }, 00:21:22.083 "auth": { 00:21:22.083 "state": "completed", 00:21:22.083 "digest": "sha384", 00:21:22.083 "dhgroup": "ffdhe6144" 00:21:22.083 } 00:21:22.083 } 00:21:22.083 ]' 00:21:22.083 08:57:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:22.083 08:57:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:22.083 08:57:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:22.083 08:57:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:21:22.083 08:57:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:22.083 08:57:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:22.083 08:57:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:22.083 08:57:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:22.344 08:57:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:22.344 08:57:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:22.914 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:23.175 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 2 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:23.175 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:23.746 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:23.746 { 00:21:23.746 "cntlid": 85, 00:21:23.746 "qid": 0, 00:21:23.746 "state": "enabled", 00:21:23.746 "thread": "nvmf_tgt_poll_group_000", 00:21:23.746 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:23.746 "listen_address": { 00:21:23.746 "trtype": "TCP", 00:21:23.746 "adrfam": "IPv4", 00:21:23.746 "traddr": "10.0.0.2", 00:21:23.746 "trsvcid": "4420" 00:21:23.746 }, 00:21:23.746 "peer_address": { 00:21:23.746 "trtype": "TCP", 00:21:23.746 "adrfam": "IPv4", 00:21:23.746 "traddr": "10.0.0.1", 00:21:23.746 "trsvcid": "36550" 00:21:23.746 }, 00:21:23.746 "auth": { 00:21:23.746 "state": "completed", 00:21:23.746 "digest": "sha384", 00:21:23.746 "dhgroup": "ffdhe6144" 00:21:23.746 } 00:21:23.746 } 00:21:23.746 ]' 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:23.746 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:24.006 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:21:24.006 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:24.006 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:24.006 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:24.006 08:57:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:24.006 08:57:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:24.006 08:57:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:24.945 08:57:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:24.945 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:24.945 08:57:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:24.945 08:57:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:24.945 08:57:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:24.945 08:57:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:24.945 08:57:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:24.945 08:57:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:21:24.945 08:57:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe6144 3 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:25.207 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:25.468 00:21:25.468 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:25.468 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:25.468 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:25.728 { 00:21:25.728 "cntlid": 87, 00:21:25.728 "qid": 0, 00:21:25.728 "state": "enabled", 00:21:25.728 "thread": "nvmf_tgt_poll_group_000", 00:21:25.728 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:25.728 "listen_address": { 00:21:25.728 "trtype": "TCP", 00:21:25.728 "adrfam": "IPv4", 00:21:25.728 "traddr": "10.0.0.2", 00:21:25.728 "trsvcid": "4420" 00:21:25.728 }, 00:21:25.728 "peer_address": { 00:21:25.728 "trtype": "TCP", 00:21:25.728 "adrfam": "IPv4", 00:21:25.728 "traddr": "10.0.0.1", 00:21:25.728 "trsvcid": "36584" 00:21:25.728 }, 00:21:25.728 "auth": { 00:21:25.728 "state": "completed", 00:21:25.728 "digest": "sha384", 00:21:25.728 "dhgroup": "ffdhe6144" 00:21:25.728 } 00:21:25.728 } 00:21:25.728 ]' 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:25.728 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:25.988 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:25.988 08:57:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:26.645 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:26.645 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:26.645 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:26.645 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:26.645 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:26.645 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:26.645 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:21:26.645 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:26.645 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:21:26.645 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 0 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:26.932 08:57:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:27.504 00:21:27.504 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:27.504 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:27.504 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:27.504 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:27.504 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:27.504 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:27.504 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:27.504 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:27.504 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:27.504 { 00:21:27.504 "cntlid": 89, 00:21:27.504 "qid": 0, 00:21:27.505 "state": "enabled", 00:21:27.505 "thread": "nvmf_tgt_poll_group_000", 00:21:27.505 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:27.505 "listen_address": { 00:21:27.505 "trtype": "TCP", 00:21:27.505 "adrfam": "IPv4", 00:21:27.505 "traddr": "10.0.0.2", 00:21:27.505 "trsvcid": "4420" 00:21:27.505 }, 00:21:27.505 "peer_address": { 00:21:27.505 "trtype": "TCP", 00:21:27.505 "adrfam": "IPv4", 00:21:27.505 "traddr": "10.0.0.1", 00:21:27.505 "trsvcid": "36600" 00:21:27.505 }, 00:21:27.505 "auth": { 00:21:27.505 "state": "completed", 00:21:27.505 "digest": "sha384", 00:21:27.505 "dhgroup": "ffdhe8192" 00:21:27.505 } 00:21:27.505 } 00:21:27.505 ]' 00:21:27.765 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:27.765 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:27.765 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:27.765 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:21:27.765 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:27.765 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:27.765 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:27.765 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:28.025 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:28.025 08:57:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:28.596 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:28.596 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:28.596 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:28.596 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:28.596 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:28.596 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:28.596 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:28.596 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:21:28.596 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 1 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:28.857 08:57:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:29.427 00:21:29.427 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:29.427 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:29.427 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:29.688 { 00:21:29.688 "cntlid": 91, 00:21:29.688 "qid": 0, 00:21:29.688 "state": "enabled", 00:21:29.688 "thread": "nvmf_tgt_poll_group_000", 00:21:29.688 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:29.688 "listen_address": { 00:21:29.688 "trtype": "TCP", 00:21:29.688 "adrfam": "IPv4", 00:21:29.688 "traddr": "10.0.0.2", 00:21:29.688 "trsvcid": "4420" 00:21:29.688 }, 00:21:29.688 "peer_address": { 00:21:29.688 "trtype": "TCP", 00:21:29.688 "adrfam": "IPv4", 00:21:29.688 "traddr": "10.0.0.1", 00:21:29.688 "trsvcid": "57944" 00:21:29.688 }, 00:21:29.688 "auth": { 00:21:29.688 "state": "completed", 00:21:29.688 "digest": "sha384", 00:21:29.688 "dhgroup": "ffdhe8192" 00:21:29.688 } 00:21:29.688 } 00:21:29.688 ]' 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:29.688 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:29.949 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:29.949 08:57:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:30.890 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 2 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:30.890 08:57:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:31.463 00:21:31.463 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:31.463 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:31.463 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:31.463 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:31.463 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:31.463 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:31.463 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:31.463 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:31.723 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:31.724 { 00:21:31.724 "cntlid": 93, 00:21:31.724 "qid": 0, 00:21:31.724 "state": "enabled", 00:21:31.724 "thread": "nvmf_tgt_poll_group_000", 00:21:31.724 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:31.724 "listen_address": { 00:21:31.724 "trtype": "TCP", 00:21:31.724 "adrfam": "IPv4", 00:21:31.724 "traddr": "10.0.0.2", 00:21:31.724 "trsvcid": "4420" 00:21:31.724 }, 00:21:31.724 "peer_address": { 00:21:31.724 "trtype": "TCP", 00:21:31.724 "adrfam": "IPv4", 00:21:31.724 "traddr": "10.0.0.1", 00:21:31.724 "trsvcid": "57954" 00:21:31.724 }, 00:21:31.724 "auth": { 00:21:31.724 "state": "completed", 00:21:31.724 "digest": "sha384", 00:21:31.724 "dhgroup": "ffdhe8192" 00:21:31.724 } 00:21:31.724 } 00:21:31.724 ]' 00:21:31.724 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:31.724 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:31.724 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:31.724 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:21:31.724 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:31.724 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:31.724 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:31.724 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:31.984 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:31.984 08:57:21 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:32.555 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:32.555 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:32.555 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:32.555 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:32.555 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:32.555 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:32.555 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:32.555 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:21:32.555 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha384 ffdhe8192 3 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha384 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:21:32.816 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:32.817 08:57:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:33.387 00:21:33.387 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:33.387 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:33.387 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:33.648 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:33.648 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:33.648 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:33.648 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:33.649 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:33.649 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:33.649 { 00:21:33.649 "cntlid": 95, 00:21:33.649 "qid": 0, 00:21:33.649 "state": "enabled", 00:21:33.649 "thread": "nvmf_tgt_poll_group_000", 00:21:33.649 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:33.649 "listen_address": { 00:21:33.649 "trtype": "TCP", 00:21:33.649 "adrfam": "IPv4", 00:21:33.649 "traddr": "10.0.0.2", 00:21:33.649 "trsvcid": "4420" 00:21:33.649 }, 00:21:33.649 "peer_address": { 00:21:33.649 "trtype": "TCP", 00:21:33.649 "adrfam": "IPv4", 00:21:33.649 "traddr": "10.0.0.1", 00:21:33.649 "trsvcid": "57994" 00:21:33.649 }, 00:21:33.649 "auth": { 00:21:33.649 "state": "completed", 00:21:33.649 "digest": "sha384", 00:21:33.649 "dhgroup": "ffdhe8192" 00:21:33.649 } 00:21:33.649 } 00:21:33.649 ]' 00:21:33.649 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:33.649 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:21:33.649 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:33.649 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:21:33.649 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:33.649 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:33.649 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:33.649 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:33.910 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:33.910 08:57:23 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:34.851 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@118 -- # for digest in "${digests[@]}" 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 0 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:34.851 08:57:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:35.112 00:21:35.112 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:35.112 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:35.112 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:35.373 { 00:21:35.373 "cntlid": 97, 00:21:35.373 "qid": 0, 00:21:35.373 "state": "enabled", 00:21:35.373 "thread": "nvmf_tgt_poll_group_000", 00:21:35.373 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:35.373 "listen_address": { 00:21:35.373 "trtype": "TCP", 00:21:35.373 "adrfam": "IPv4", 00:21:35.373 "traddr": "10.0.0.2", 00:21:35.373 "trsvcid": "4420" 00:21:35.373 }, 00:21:35.373 "peer_address": { 00:21:35.373 "trtype": "TCP", 00:21:35.373 "adrfam": "IPv4", 00:21:35.373 "traddr": "10.0.0.1", 00:21:35.373 "trsvcid": "58006" 00:21:35.373 }, 00:21:35.373 "auth": { 00:21:35.373 "state": "completed", 00:21:35.373 "digest": "sha512", 00:21:35.373 "dhgroup": "null" 00:21:35.373 } 00:21:35.373 } 00:21:35.373 ]' 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:35.373 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:35.634 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:35.634 08:57:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:36.574 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 1 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:36.574 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:36.833 00:21:36.833 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:36.833 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:36.833 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:37.094 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:37.094 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:37.094 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:37.094 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:37.094 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:37.094 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:37.094 { 00:21:37.094 "cntlid": 99, 00:21:37.094 "qid": 0, 00:21:37.094 "state": "enabled", 00:21:37.094 "thread": "nvmf_tgt_poll_group_000", 00:21:37.094 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:37.094 "listen_address": { 00:21:37.094 "trtype": "TCP", 00:21:37.094 "adrfam": "IPv4", 00:21:37.094 "traddr": "10.0.0.2", 00:21:37.094 "trsvcid": "4420" 00:21:37.094 }, 00:21:37.094 "peer_address": { 00:21:37.094 "trtype": "TCP", 00:21:37.094 "adrfam": "IPv4", 00:21:37.094 "traddr": "10.0.0.1", 00:21:37.094 "trsvcid": "58028" 00:21:37.094 }, 00:21:37.094 "auth": { 00:21:37.094 "state": "completed", 00:21:37.094 "digest": "sha512", 00:21:37.094 "dhgroup": "null" 00:21:37.094 } 00:21:37.094 } 00:21:37.094 ]' 00:21:37.094 08:57:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:37.094 08:57:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:37.094 08:57:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:37.094 08:57:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:21:37.094 08:57:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:37.094 08:57:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:37.094 08:57:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:37.094 08:57:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:37.354 08:57:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:37.354 08:57:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:37.931 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:38.191 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 2 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:38.191 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:38.452 00:21:38.452 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:38.452 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:38.452 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:38.712 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:38.712 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:38.712 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:38.712 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:38.712 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:38.712 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:38.712 { 00:21:38.712 "cntlid": 101, 00:21:38.712 "qid": 0, 00:21:38.712 "state": "enabled", 00:21:38.712 "thread": "nvmf_tgt_poll_group_000", 00:21:38.712 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:38.712 "listen_address": { 00:21:38.712 "trtype": "TCP", 00:21:38.712 "adrfam": "IPv4", 00:21:38.712 "traddr": "10.0.0.2", 00:21:38.712 "trsvcid": "4420" 00:21:38.712 }, 00:21:38.712 "peer_address": { 00:21:38.712 "trtype": "TCP", 00:21:38.712 "adrfam": "IPv4", 00:21:38.712 "traddr": "10.0.0.1", 00:21:38.712 "trsvcid": "53148" 00:21:38.712 }, 00:21:38.712 "auth": { 00:21:38.712 "state": "completed", 00:21:38.712 "digest": "sha512", 00:21:38.712 "dhgroup": "null" 00:21:38.712 } 00:21:38.712 } 00:21:38.712 ]' 00:21:38.712 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:38.712 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:38.712 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:38.712 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:21:38.713 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:38.713 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:38.713 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:38.713 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:38.974 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:38.974 08:57:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:39.915 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 null 3 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=null 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:39.915 08:57:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:40.176 00:21:40.176 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:40.176 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:40.176 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:40.436 { 00:21:40.436 "cntlid": 103, 00:21:40.436 "qid": 0, 00:21:40.436 "state": "enabled", 00:21:40.436 "thread": "nvmf_tgt_poll_group_000", 00:21:40.436 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:40.436 "listen_address": { 00:21:40.436 "trtype": "TCP", 00:21:40.436 "adrfam": "IPv4", 00:21:40.436 "traddr": "10.0.0.2", 00:21:40.436 "trsvcid": "4420" 00:21:40.436 }, 00:21:40.436 "peer_address": { 00:21:40.436 "trtype": "TCP", 00:21:40.436 "adrfam": "IPv4", 00:21:40.436 "traddr": "10.0.0.1", 00:21:40.436 "trsvcid": "53178" 00:21:40.436 }, 00:21:40.436 "auth": { 00:21:40.436 "state": "completed", 00:21:40.436 "digest": "sha512", 00:21:40.436 "dhgroup": "null" 00:21:40.436 } 00:21:40.436 } 00:21:40.436 ]' 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ null == \n\u\l\l ]] 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:40.436 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:40.697 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:40.697 08:57:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:41.638 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 0 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:41.638 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:41.899 00:21:41.899 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:41.899 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:41.899 08:57:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:42.159 { 00:21:42.159 "cntlid": 105, 00:21:42.159 "qid": 0, 00:21:42.159 "state": "enabled", 00:21:42.159 "thread": "nvmf_tgt_poll_group_000", 00:21:42.159 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:42.159 "listen_address": { 00:21:42.159 "trtype": "TCP", 00:21:42.159 "adrfam": "IPv4", 00:21:42.159 "traddr": "10.0.0.2", 00:21:42.159 "trsvcid": "4420" 00:21:42.159 }, 00:21:42.159 "peer_address": { 00:21:42.159 "trtype": "TCP", 00:21:42.159 "adrfam": "IPv4", 00:21:42.159 "traddr": "10.0.0.1", 00:21:42.159 "trsvcid": "53208" 00:21:42.159 }, 00:21:42.159 "auth": { 00:21:42.159 "state": "completed", 00:21:42.159 "digest": "sha512", 00:21:42.159 "dhgroup": "ffdhe2048" 00:21:42.159 } 00:21:42.159 } 00:21:42.159 ]' 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:42.159 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:42.420 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:42.420 08:57:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:43.361 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:43.361 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 1 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:43.362 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:43.623 00:21:43.623 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:43.623 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:43.623 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:43.884 { 00:21:43.884 "cntlid": 107, 00:21:43.884 "qid": 0, 00:21:43.884 "state": "enabled", 00:21:43.884 "thread": "nvmf_tgt_poll_group_000", 00:21:43.884 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:43.884 "listen_address": { 00:21:43.884 "trtype": "TCP", 00:21:43.884 "adrfam": "IPv4", 00:21:43.884 "traddr": "10.0.0.2", 00:21:43.884 "trsvcid": "4420" 00:21:43.884 }, 00:21:43.884 "peer_address": { 00:21:43.884 "trtype": "TCP", 00:21:43.884 "adrfam": "IPv4", 00:21:43.884 "traddr": "10.0.0.1", 00:21:43.884 "trsvcid": "53244" 00:21:43.884 }, 00:21:43.884 "auth": { 00:21:43.884 "state": "completed", 00:21:43.884 "digest": "sha512", 00:21:43.884 "dhgroup": "ffdhe2048" 00:21:43.884 } 00:21:43.884 } 00:21:43.884 ]' 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:43.884 08:57:33 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:44.145 08:57:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:44.145 08:57:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:44.716 08:57:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:44.716 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:44.716 08:57:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:44.716 08:57:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:44.716 08:57:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:44.716 08:57:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:44.716 08:57:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:44.716 08:57:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:21:44.716 08:57:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 2 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:44.977 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:45.237 00:21:45.237 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:45.237 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:45.237 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:45.497 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:45.497 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:45.497 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:45.497 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:45.497 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:45.498 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:45.498 { 00:21:45.498 "cntlid": 109, 00:21:45.498 "qid": 0, 00:21:45.498 "state": "enabled", 00:21:45.498 "thread": "nvmf_tgt_poll_group_000", 00:21:45.498 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:45.498 "listen_address": { 00:21:45.498 "trtype": "TCP", 00:21:45.498 "adrfam": "IPv4", 00:21:45.498 "traddr": "10.0.0.2", 00:21:45.498 "trsvcid": "4420" 00:21:45.498 }, 00:21:45.498 "peer_address": { 00:21:45.498 "trtype": "TCP", 00:21:45.498 "adrfam": "IPv4", 00:21:45.498 "traddr": "10.0.0.1", 00:21:45.498 "trsvcid": "53276" 00:21:45.498 }, 00:21:45.498 "auth": { 00:21:45.498 "state": "completed", 00:21:45.498 "digest": "sha512", 00:21:45.498 "dhgroup": "ffdhe2048" 00:21:45.498 } 00:21:45.498 } 00:21:45.498 ]' 00:21:45.498 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:45.498 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:45.498 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:45.498 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:21:45.498 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:45.498 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:45.498 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:45.498 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:45.758 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:45.758 08:57:35 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:46.702 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe2048 3 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe2048 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:46.702 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:46.963 00:21:46.963 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:46.963 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:46.963 08:57:36 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:47.223 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:47.223 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:47.223 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:47.223 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:47.224 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:47.224 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:47.224 { 00:21:47.224 "cntlid": 111, 00:21:47.224 "qid": 0, 00:21:47.224 "state": "enabled", 00:21:47.224 "thread": "nvmf_tgt_poll_group_000", 00:21:47.224 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:47.224 "listen_address": { 00:21:47.224 "trtype": "TCP", 00:21:47.224 "adrfam": "IPv4", 00:21:47.224 "traddr": "10.0.0.2", 00:21:47.224 "trsvcid": "4420" 00:21:47.224 }, 00:21:47.224 "peer_address": { 00:21:47.224 "trtype": "TCP", 00:21:47.224 "adrfam": "IPv4", 00:21:47.224 "traddr": "10.0.0.1", 00:21:47.224 "trsvcid": "53298" 00:21:47.224 }, 00:21:47.224 "auth": { 00:21:47.224 "state": "completed", 00:21:47.224 "digest": "sha512", 00:21:47.224 "dhgroup": "ffdhe2048" 00:21:47.224 } 00:21:47.224 } 00:21:47.224 ]' 00:21:47.224 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:47.224 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:47.224 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:47.224 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:21:47.224 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:47.224 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:47.224 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:47.224 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:47.485 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:47.485 08:57:37 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:48.058 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:48.319 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 0 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:48.319 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:48.581 00:21:48.581 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:48.581 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:48.581 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:48.842 { 00:21:48.842 "cntlid": 113, 00:21:48.842 "qid": 0, 00:21:48.842 "state": "enabled", 00:21:48.842 "thread": "nvmf_tgt_poll_group_000", 00:21:48.842 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:48.842 "listen_address": { 00:21:48.842 "trtype": "TCP", 00:21:48.842 "adrfam": "IPv4", 00:21:48.842 "traddr": "10.0.0.2", 00:21:48.842 "trsvcid": "4420" 00:21:48.842 }, 00:21:48.842 "peer_address": { 00:21:48.842 "trtype": "TCP", 00:21:48.842 "adrfam": "IPv4", 00:21:48.842 "traddr": "10.0.0.1", 00:21:48.842 "trsvcid": "37534" 00:21:48.842 }, 00:21:48.842 "auth": { 00:21:48.842 "state": "completed", 00:21:48.842 "digest": "sha512", 00:21:48.842 "dhgroup": "ffdhe3072" 00:21:48.842 } 00:21:48.842 } 00:21:48.842 ]' 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:21:48.842 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:49.102 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:49.102 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:49.102 08:57:38 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:49.102 08:57:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:49.103 08:57:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:50.047 08:57:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:50.047 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:50.047 08:57:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:50.047 08:57:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:50.047 08:57:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:50.047 08:57:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:50.047 08:57:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:50.047 08:57:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:21:50.047 08:57:39 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 1 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:50.047 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:50.307 00:21:50.307 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:50.307 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:50.307 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:50.568 { 00:21:50.568 "cntlid": 115, 00:21:50.568 "qid": 0, 00:21:50.568 "state": "enabled", 00:21:50.568 "thread": "nvmf_tgt_poll_group_000", 00:21:50.568 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:50.568 "listen_address": { 00:21:50.568 "trtype": "TCP", 00:21:50.568 "adrfam": "IPv4", 00:21:50.568 "traddr": "10.0.0.2", 00:21:50.568 "trsvcid": "4420" 00:21:50.568 }, 00:21:50.568 "peer_address": { 00:21:50.568 "trtype": "TCP", 00:21:50.568 "adrfam": "IPv4", 00:21:50.568 "traddr": "10.0.0.1", 00:21:50.568 "trsvcid": "37560" 00:21:50.568 }, 00:21:50.568 "auth": { 00:21:50.568 "state": "completed", 00:21:50.568 "digest": "sha512", 00:21:50.568 "dhgroup": "ffdhe3072" 00:21:50.568 } 00:21:50.568 } 00:21:50.568 ]' 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:50.568 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:50.828 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:50.828 08:57:40 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:51.770 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 2 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:51.770 08:57:41 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:52.031 00:21:52.031 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:52.031 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:52.031 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:52.292 { 00:21:52.292 "cntlid": 117, 00:21:52.292 "qid": 0, 00:21:52.292 "state": "enabled", 00:21:52.292 "thread": "nvmf_tgt_poll_group_000", 00:21:52.292 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:52.292 "listen_address": { 00:21:52.292 "trtype": "TCP", 00:21:52.292 "adrfam": "IPv4", 00:21:52.292 "traddr": "10.0.0.2", 00:21:52.292 "trsvcid": "4420" 00:21:52.292 }, 00:21:52.292 "peer_address": { 00:21:52.292 "trtype": "TCP", 00:21:52.292 "adrfam": "IPv4", 00:21:52.292 "traddr": "10.0.0.1", 00:21:52.292 "trsvcid": "37588" 00:21:52.292 }, 00:21:52.292 "auth": { 00:21:52.292 "state": "completed", 00:21:52.292 "digest": "sha512", 00:21:52.292 "dhgroup": "ffdhe3072" 00:21:52.292 } 00:21:52.292 } 00:21:52.292 ]' 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:52.292 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:52.552 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:52.552 08:57:42 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:53.495 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe3072 3 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe3072 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:53.495 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:21:53.755 00:21:53.755 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:53.755 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:53.755 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:54.016 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:54.016 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:54.016 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:54.016 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:54.016 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:54.016 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:54.016 { 00:21:54.016 "cntlid": 119, 00:21:54.016 "qid": 0, 00:21:54.016 "state": "enabled", 00:21:54.016 "thread": "nvmf_tgt_poll_group_000", 00:21:54.016 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:54.016 "listen_address": { 00:21:54.016 "trtype": "TCP", 00:21:54.016 "adrfam": "IPv4", 00:21:54.016 "traddr": "10.0.0.2", 00:21:54.016 "trsvcid": "4420" 00:21:54.016 }, 00:21:54.016 "peer_address": { 00:21:54.016 "trtype": "TCP", 00:21:54.016 "adrfam": "IPv4", 00:21:54.016 "traddr": "10.0.0.1", 00:21:54.016 "trsvcid": "37608" 00:21:54.016 }, 00:21:54.016 "auth": { 00:21:54.016 "state": "completed", 00:21:54.016 "digest": "sha512", 00:21:54.016 "dhgroup": "ffdhe3072" 00:21:54.016 } 00:21:54.016 } 00:21:54.016 ]' 00:21:54.016 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:54.016 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:54.016 08:57:43 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:54.016 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:21:54.016 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:54.016 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:54.016 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:54.016 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:54.277 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:54.277 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:21:55.219 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:55.219 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:55.219 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:55.219 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:55.220 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:55.220 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:55.220 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:21:55.220 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:55.220 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:21:55.220 08:57:44 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 0 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:55.220 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:21:55.480 00:21:55.480 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:55.480 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:55.480 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:55.741 { 00:21:55.741 "cntlid": 121, 00:21:55.741 "qid": 0, 00:21:55.741 "state": "enabled", 00:21:55.741 "thread": "nvmf_tgt_poll_group_000", 00:21:55.741 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:55.741 "listen_address": { 00:21:55.741 "trtype": "TCP", 00:21:55.741 "adrfam": "IPv4", 00:21:55.741 "traddr": "10.0.0.2", 00:21:55.741 "trsvcid": "4420" 00:21:55.741 }, 00:21:55.741 "peer_address": { 00:21:55.741 "trtype": "TCP", 00:21:55.741 "adrfam": "IPv4", 00:21:55.741 "traddr": "10.0.0.1", 00:21:55.741 "trsvcid": "37634" 00:21:55.741 }, 00:21:55.741 "auth": { 00:21:55.741 "state": "completed", 00:21:55.741 "digest": "sha512", 00:21:55.741 "dhgroup": "ffdhe4096" 00:21:55.741 } 00:21:55.741 } 00:21:55.741 ]' 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:55.741 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:56.001 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:56.001 08:57:45 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:56.944 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 1 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:56.944 08:57:46 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:21:57.206 00:21:57.206 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:57.206 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:57.206 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:57.466 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:57.466 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:57.466 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:57.466 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:57.466 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:57.466 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:57.466 { 00:21:57.466 "cntlid": 123, 00:21:57.466 "qid": 0, 00:21:57.466 "state": "enabled", 00:21:57.466 "thread": "nvmf_tgt_poll_group_000", 00:21:57.467 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:57.467 "listen_address": { 00:21:57.467 "trtype": "TCP", 00:21:57.467 "adrfam": "IPv4", 00:21:57.467 "traddr": "10.0.0.2", 00:21:57.467 "trsvcid": "4420" 00:21:57.467 }, 00:21:57.467 "peer_address": { 00:21:57.467 "trtype": "TCP", 00:21:57.467 "adrfam": "IPv4", 00:21:57.467 "traddr": "10.0.0.1", 00:21:57.467 "trsvcid": "37658" 00:21:57.467 }, 00:21:57.467 "auth": { 00:21:57.467 "state": "completed", 00:21:57.467 "digest": "sha512", 00:21:57.467 "dhgroup": "ffdhe4096" 00:21:57.467 } 00:21:57.467 } 00:21:57.467 ]' 00:21:57.467 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:57.467 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:57.467 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:57.467 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:21:57.467 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:57.467 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:57.467 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:57.467 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:57.737 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:57.737 08:57:47 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:21:58.679 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 2 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:58.679 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:21:58.939 00:21:58.939 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:21:58.939 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:21:58.939 08:57:48 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:21:59.199 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:21:59.199 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:21:59.199 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:21:59.199 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:21:59.199 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:21:59.199 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:21:59.199 { 00:21:59.199 "cntlid": 125, 00:21:59.199 "qid": 0, 00:21:59.199 "state": "enabled", 00:21:59.199 "thread": "nvmf_tgt_poll_group_000", 00:21:59.199 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:21:59.199 "listen_address": { 00:21:59.199 "trtype": "TCP", 00:21:59.199 "adrfam": "IPv4", 00:21:59.199 "traddr": "10.0.0.2", 00:21:59.199 "trsvcid": "4420" 00:21:59.199 }, 00:21:59.199 "peer_address": { 00:21:59.199 "trtype": "TCP", 00:21:59.199 "adrfam": "IPv4", 00:21:59.199 "traddr": "10.0.0.1", 00:21:59.199 "trsvcid": "49748" 00:21:59.199 }, 00:21:59.199 "auth": { 00:21:59.199 "state": "completed", 00:21:59.199 "digest": "sha512", 00:21:59.199 "dhgroup": "ffdhe4096" 00:21:59.199 } 00:21:59.199 } 00:21:59.199 ]' 00:21:59.199 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:21:59.199 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:21:59.200 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:21:59.200 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:21:59.200 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:21:59.200 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:21:59.200 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:21:59.200 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:21:59.459 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:21:59.459 08:57:49 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:00.401 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe4096 3 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe4096 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:00.401 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:00.661 00:22:00.661 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:00.661 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:00.661 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:00.922 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:00.922 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:00.922 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:00.922 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:00.922 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:00.922 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:00.922 { 00:22:00.922 "cntlid": 127, 00:22:00.922 "qid": 0, 00:22:00.922 "state": "enabled", 00:22:00.922 "thread": "nvmf_tgt_poll_group_000", 00:22:00.923 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:00.923 "listen_address": { 00:22:00.923 "trtype": "TCP", 00:22:00.923 "adrfam": "IPv4", 00:22:00.923 "traddr": "10.0.0.2", 00:22:00.923 "trsvcid": "4420" 00:22:00.923 }, 00:22:00.923 "peer_address": { 00:22:00.923 "trtype": "TCP", 00:22:00.923 "adrfam": "IPv4", 00:22:00.923 "traddr": "10.0.0.1", 00:22:00.923 "trsvcid": "49786" 00:22:00.923 }, 00:22:00.923 "auth": { 00:22:00.923 "state": "completed", 00:22:00.923 "digest": "sha512", 00:22:00.923 "dhgroup": "ffdhe4096" 00:22:00.923 } 00:22:00.923 } 00:22:00.923 ]' 00:22:00.923 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:00.923 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:00.923 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:00.923 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:22:00.923 08:57:50 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:00.923 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:00.923 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:00.923 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:01.184 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:22:01.184 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:22:02.124 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:02.125 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:02.125 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:02.125 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:02.125 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:02.125 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:02.125 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:22:02.125 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:02.125 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:22:02.125 08:57:51 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 0 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:02.125 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:02.385 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:02.645 { 00:22:02.645 "cntlid": 129, 00:22:02.645 "qid": 0, 00:22:02.645 "state": "enabled", 00:22:02.645 "thread": "nvmf_tgt_poll_group_000", 00:22:02.645 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:02.645 "listen_address": { 00:22:02.645 "trtype": "TCP", 00:22:02.645 "adrfam": "IPv4", 00:22:02.645 "traddr": "10.0.0.2", 00:22:02.645 "trsvcid": "4420" 00:22:02.645 }, 00:22:02.645 "peer_address": { 00:22:02.645 "trtype": "TCP", 00:22:02.645 "adrfam": "IPv4", 00:22:02.645 "traddr": "10.0.0.1", 00:22:02.645 "trsvcid": "49812" 00:22:02.645 }, 00:22:02.645 "auth": { 00:22:02.645 "state": "completed", 00:22:02.645 "digest": "sha512", 00:22:02.645 "dhgroup": "ffdhe6144" 00:22:02.645 } 00:22:02.645 } 00:22:02.645 ]' 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:02.645 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:02.906 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:22:02.906 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:02.906 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:02.906 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:02.906 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:02.906 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:22:02.906 08:57:52 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:03.848 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 1 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:03.848 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:04.109 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:04.109 08:57:53 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:04.369 00:22:04.369 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:04.369 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:04.369 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:04.629 { 00:22:04.629 "cntlid": 131, 00:22:04.629 "qid": 0, 00:22:04.629 "state": "enabled", 00:22:04.629 "thread": "nvmf_tgt_poll_group_000", 00:22:04.629 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:04.629 "listen_address": { 00:22:04.629 "trtype": "TCP", 00:22:04.629 "adrfam": "IPv4", 00:22:04.629 "traddr": "10.0.0.2", 00:22:04.629 "trsvcid": "4420" 00:22:04.629 }, 00:22:04.629 "peer_address": { 00:22:04.629 "trtype": "TCP", 00:22:04.629 "adrfam": "IPv4", 00:22:04.629 "traddr": "10.0.0.1", 00:22:04.629 "trsvcid": "49838" 00:22:04.629 }, 00:22:04.629 "auth": { 00:22:04.629 "state": "completed", 00:22:04.629 "digest": "sha512", 00:22:04.629 "dhgroup": "ffdhe6144" 00:22:04.629 } 00:22:04.629 } 00:22:04.629 ]' 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:04.629 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:04.890 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:22:04.890 08:57:54 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:22:05.461 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:05.461 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:05.461 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:05.461 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:05.461 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:05.461 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:05.461 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:05.461 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:22:05.461 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 2 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:05.722 08:57:55 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:06.375 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:06.375 { 00:22:06.375 "cntlid": 133, 00:22:06.375 "qid": 0, 00:22:06.375 "state": "enabled", 00:22:06.375 "thread": "nvmf_tgt_poll_group_000", 00:22:06.375 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:06.375 "listen_address": { 00:22:06.375 "trtype": "TCP", 00:22:06.375 "adrfam": "IPv4", 00:22:06.375 "traddr": "10.0.0.2", 00:22:06.375 "trsvcid": "4420" 00:22:06.375 }, 00:22:06.375 "peer_address": { 00:22:06.375 "trtype": "TCP", 00:22:06.375 "adrfam": "IPv4", 00:22:06.375 "traddr": "10.0.0.1", 00:22:06.375 "trsvcid": "49864" 00:22:06.375 }, 00:22:06.375 "auth": { 00:22:06.375 "state": "completed", 00:22:06.375 "digest": "sha512", 00:22:06.375 "dhgroup": "ffdhe6144" 00:22:06.375 } 00:22:06.375 } 00:22:06.375 ]' 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:06.375 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:06.657 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:22:06.657 08:57:56 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:07.608 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe6144 3 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe6144 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:07.608 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:07.869 00:22:08.129 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:08.129 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:08.129 08:57:57 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:08.129 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:08.129 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:08.129 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:08.129 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:08.129 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:08.129 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:08.129 { 00:22:08.129 "cntlid": 135, 00:22:08.129 "qid": 0, 00:22:08.129 "state": "enabled", 00:22:08.129 "thread": "nvmf_tgt_poll_group_000", 00:22:08.129 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:08.129 "listen_address": { 00:22:08.129 "trtype": "TCP", 00:22:08.129 "adrfam": "IPv4", 00:22:08.129 "traddr": "10.0.0.2", 00:22:08.129 "trsvcid": "4420" 00:22:08.129 }, 00:22:08.129 "peer_address": { 00:22:08.129 "trtype": "TCP", 00:22:08.129 "adrfam": "IPv4", 00:22:08.129 "traddr": "10.0.0.1", 00:22:08.129 "trsvcid": "49900" 00:22:08.129 }, 00:22:08.129 "auth": { 00:22:08.129 "state": "completed", 00:22:08.129 "digest": "sha512", 00:22:08.129 "dhgroup": "ffdhe6144" 00:22:08.129 } 00:22:08.129 } 00:22:08.129 ]' 00:22:08.129 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:08.129 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:08.129 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:08.390 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:22:08.390 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:08.390 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:08.390 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:08.390 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:08.390 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:22:08.390 08:57:58 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:09.331 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@119 -- # for dhgroup in "${dhgroups[@]}" 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 0 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:22:09.331 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:22:09.593 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:09.593 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:09.593 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:09.593 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:09.593 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:09.593 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:09.593 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:09.593 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:10.162 00:22:10.162 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:10.162 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:10.162 08:57:59 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:10.162 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:10.162 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:10.162 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:10.162 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:10.162 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:10.162 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:10.162 { 00:22:10.162 "cntlid": 137, 00:22:10.162 "qid": 0, 00:22:10.162 "state": "enabled", 00:22:10.162 "thread": "nvmf_tgt_poll_group_000", 00:22:10.162 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:10.162 "listen_address": { 00:22:10.162 "trtype": "TCP", 00:22:10.163 "adrfam": "IPv4", 00:22:10.163 "traddr": "10.0.0.2", 00:22:10.163 "trsvcid": "4420" 00:22:10.163 }, 00:22:10.163 "peer_address": { 00:22:10.163 "trtype": "TCP", 00:22:10.163 "adrfam": "IPv4", 00:22:10.163 "traddr": "10.0.0.1", 00:22:10.163 "trsvcid": "43246" 00:22:10.163 }, 00:22:10.163 "auth": { 00:22:10.163 "state": "completed", 00:22:10.163 "digest": "sha512", 00:22:10.163 "dhgroup": "ffdhe8192" 00:22:10.163 } 00:22:10.163 } 00:22:10.163 ]' 00:22:10.163 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:10.163 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:10.163 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:10.424 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:22:10.424 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:10.424 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:10.424 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:10.424 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:10.424 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:22:10.424 08:58:00 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:22:11.366 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:11.366 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:11.366 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:11.366 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:11.366 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:11.366 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:11.366 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:11.366 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:22:11.366 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 1 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key1 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:11.626 08:58:01 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:12.196 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:12.196 { 00:22:12.196 "cntlid": 139, 00:22:12.196 "qid": 0, 00:22:12.196 "state": "enabled", 00:22:12.196 "thread": "nvmf_tgt_poll_group_000", 00:22:12.196 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:12.196 "listen_address": { 00:22:12.196 "trtype": "TCP", 00:22:12.196 "adrfam": "IPv4", 00:22:12.196 "traddr": "10.0.0.2", 00:22:12.196 "trsvcid": "4420" 00:22:12.196 }, 00:22:12.196 "peer_address": { 00:22:12.196 "trtype": "TCP", 00:22:12.196 "adrfam": "IPv4", 00:22:12.196 "traddr": "10.0.0.1", 00:22:12.196 "trsvcid": "43276" 00:22:12.196 }, 00:22:12.196 "auth": { 00:22:12.196 "state": "completed", 00:22:12.196 "digest": "sha512", 00:22:12.196 "dhgroup": "ffdhe8192" 00:22:12.196 } 00:22:12.196 } 00:22:12.196 ]' 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:12.196 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:12.456 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:22:12.456 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:12.456 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:12.456 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:12.456 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:12.456 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:22:12.456 08:58:02 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: --dhchap-ctrl-secret DHHC-1:02:NTI0MjI1NjkxZDhhYzlmMmU4MmZlYmQzODY5OTJjYWRjMzJkZTg3MzQyMzBkMTcz2u6dLg==: 00:22:13.397 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:13.398 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 2 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key2 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:13.398 08:58:03 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:22:13.969 00:22:13.969 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:13.969 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:13.969 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:14.228 { 00:22:14.228 "cntlid": 141, 00:22:14.228 "qid": 0, 00:22:14.228 "state": "enabled", 00:22:14.228 "thread": "nvmf_tgt_poll_group_000", 00:22:14.228 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:14.228 "listen_address": { 00:22:14.228 "trtype": "TCP", 00:22:14.228 "adrfam": "IPv4", 00:22:14.228 "traddr": "10.0.0.2", 00:22:14.228 "trsvcid": "4420" 00:22:14.228 }, 00:22:14.228 "peer_address": { 00:22:14.228 "trtype": "TCP", 00:22:14.228 "adrfam": "IPv4", 00:22:14.228 "traddr": "10.0.0.1", 00:22:14.228 "trsvcid": "43294" 00:22:14.228 }, 00:22:14.228 "auth": { 00:22:14.228 "state": "completed", 00:22:14.228 "digest": "sha512", 00:22:14.228 "dhgroup": "ffdhe8192" 00:22:14.228 } 00:22:14.228 } 00:22:14.228 ]' 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:14.228 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:14.229 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:14.229 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:14.489 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:22:14.489 08:58:04 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:01:YmM5OWZlMzk4OGJiMGZmYzMzZTQ4NmMzYTY0ZGRlNTerH+oL: 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:15.431 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@120 -- # for keyid in "${!keys[@]}" 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@121 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@123 -- # connect_authenticate sha512 ffdhe8192 3 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:15.431 08:58:05 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:16.002 00:22:16.002 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:16.002 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:16.002 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:16.263 { 00:22:16.263 "cntlid": 143, 00:22:16.263 "qid": 0, 00:22:16.263 "state": "enabled", 00:22:16.263 "thread": "nvmf_tgt_poll_group_000", 00:22:16.263 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:16.263 "listen_address": { 00:22:16.263 "trtype": "TCP", 00:22:16.263 "adrfam": "IPv4", 00:22:16.263 "traddr": "10.0.0.2", 00:22:16.263 "trsvcid": "4420" 00:22:16.263 }, 00:22:16.263 "peer_address": { 00:22:16.263 "trtype": "TCP", 00:22:16.263 "adrfam": "IPv4", 00:22:16.263 "traddr": "10.0.0.1", 00:22:16.263 "trsvcid": "43324" 00:22:16.263 }, 00:22:16.263 "auth": { 00:22:16.263 "state": "completed", 00:22:16.263 "digest": "sha512", 00:22:16.263 "dhgroup": "ffdhe8192" 00:22:16.263 } 00:22:16.263 } 00:22:16.263 ]' 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:16.263 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:16.524 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:22:16.524 08:58:06 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:17.467 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # IFS=, 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@130 -- # printf %s sha256,sha384,sha512 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # IFS=, 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@130 -- # printf %s null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@129 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@141 -- # connect_authenticate sha512 ffdhe8192 0 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key0 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:17.467 08:58:07 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:22:18.037 00:22:18.037 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:18.037 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:18.037 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:18.297 { 00:22:18.297 "cntlid": 145, 00:22:18.297 "qid": 0, 00:22:18.297 "state": "enabled", 00:22:18.297 "thread": "nvmf_tgt_poll_group_000", 00:22:18.297 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:18.297 "listen_address": { 00:22:18.297 "trtype": "TCP", 00:22:18.297 "adrfam": "IPv4", 00:22:18.297 "traddr": "10.0.0.2", 00:22:18.297 "trsvcid": "4420" 00:22:18.297 }, 00:22:18.297 "peer_address": { 00:22:18.297 "trtype": "TCP", 00:22:18.297 "adrfam": "IPv4", 00:22:18.297 "traddr": "10.0.0.1", 00:22:18.297 "trsvcid": "43356" 00:22:18.297 }, 00:22:18.297 "auth": { 00:22:18.297 "state": "completed", 00:22:18.297 "digest": "sha512", 00:22:18.297 "dhgroup": "ffdhe8192" 00:22:18.297 } 00:22:18.297 } 00:22:18.297 ]' 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:18.297 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:18.298 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:18.558 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:22:18.558 08:58:08 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:00:OTFjNWQ5YzdjZTI3OWUyM2QyMTc2M2Y2ODYyMDllNjlmMzYyMzdjNmVmODUyN2Y5L2Ij8Q==: --dhchap-ctrl-secret DHHC-1:03:YTUyMmVjYWJkZjM0OGMwY2M0Mjc4YjAxY2Q1Nzc4MGJmY2M5NzYzYjdhZDI3ZWEzYjFiMGRkZDg0OTRiMDAzN9kkudo=: 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:19.500 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@144 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@145 -- # NOT bdev_connect -b nvme0 --dhchap-key key2 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key2 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key2 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 00:22:19.500 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 00:22:19.760 request: 00:22:19.760 { 00:22:19.760 "name": "nvme0", 00:22:19.760 "trtype": "tcp", 00:22:19.760 "traddr": "10.0.0.2", 00:22:19.760 "adrfam": "ipv4", 00:22:19.760 "trsvcid": "4420", 00:22:19.760 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:22:19.760 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:19.760 "prchk_reftag": false, 00:22:19.760 "prchk_guard": false, 00:22:19.760 "hdgst": false, 00:22:19.760 "ddgst": false, 00:22:19.760 "dhchap_key": "key2", 00:22:19.760 "allow_unrecognized_csi": false, 00:22:19.760 "method": "bdev_nvme_attach_controller", 00:22:19.760 "req_id": 1 00:22:19.760 } 00:22:19.760 Got JSON-RPC error response 00:22:19.760 response: 00:22:19.760 { 00:22:19.760 "code": -5, 00:22:19.760 "message": "Input/output error" 00:22:19.760 } 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@146 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@149 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@150 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:22:19.760 08:58:09 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:22:20.330 request: 00:22:20.330 { 00:22:20.330 "name": "nvme0", 00:22:20.330 "trtype": "tcp", 00:22:20.330 "traddr": "10.0.0.2", 00:22:20.330 "adrfam": "ipv4", 00:22:20.330 "trsvcid": "4420", 00:22:20.330 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:22:20.330 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:20.330 "prchk_reftag": false, 00:22:20.330 "prchk_guard": false, 00:22:20.330 "hdgst": false, 00:22:20.330 "ddgst": false, 00:22:20.330 "dhchap_key": "key1", 00:22:20.330 "dhchap_ctrlr_key": "ckey2", 00:22:20.330 "allow_unrecognized_csi": false, 00:22:20.330 "method": "bdev_nvme_attach_controller", 00:22:20.330 "req_id": 1 00:22:20.330 } 00:22:20.330 Got JSON-RPC error response 00:22:20.330 response: 00:22:20.330 { 00:22:20.330 "code": -5, 00:22:20.330 "message": "Input/output error" 00:22:20.330 } 00:22:20.330 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:22:20.330 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:20.330 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:20.330 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:20.330 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@151 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:20.330 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:20.330 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:20.330 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:20.330 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@154 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@155 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:20.331 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:22:20.901 request: 00:22:20.901 { 00:22:20.901 "name": "nvme0", 00:22:20.901 "trtype": "tcp", 00:22:20.901 "traddr": "10.0.0.2", 00:22:20.901 "adrfam": "ipv4", 00:22:20.901 "trsvcid": "4420", 00:22:20.901 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:22:20.901 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:20.901 "prchk_reftag": false, 00:22:20.901 "prchk_guard": false, 00:22:20.901 "hdgst": false, 00:22:20.901 "ddgst": false, 00:22:20.901 "dhchap_key": "key1", 00:22:20.901 "dhchap_ctrlr_key": "ckey1", 00:22:20.901 "allow_unrecognized_csi": false, 00:22:20.901 "method": "bdev_nvme_attach_controller", 00:22:20.901 "req_id": 1 00:22:20.901 } 00:22:20.901 Got JSON-RPC error response 00:22:20.901 response: 00:22:20.901 { 00:22:20.901 "code": -5, 00:22:20.901 "message": "Input/output error" 00:22:20.901 } 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@156 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@159 -- # killprocess 1866650 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@950 -- # '[' -z 1866650 ']' 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # kill -0 1866650 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # uname 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1866650 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1866650' 00:22:20.901 killing process with pid 1866650 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@969 -- # kill 1866650 00:22:20.901 08:58:10 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@974 -- # wait 1866650 00:22:21.161 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@160 -- # nvmfappstart --wait-for-rpc -L nvmf_auth 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@724 -- # xtrace_disable 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@507 -- # nvmfpid=1894314 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc -L nvmf_auth 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@508 -- # waitforlisten 1894314 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@831 -- # '[' -z 1894314 ']' 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:21.162 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # return 0 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@730 -- # xtrace_disable 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@161 -- # trap 'dumplogs; cleanup' SIGINT SIGTERM EXIT 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@163 -- # waitforlisten 1894314 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@831 -- # '[' -z 1894314 ']' 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:22.102 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:22.102 08:58:11 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.102 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:22.102 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@864 -- # return 0 00:22:22.102 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@164 -- # rpc_cmd 00:22:22.102 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.102 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.363 null0 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.2c1 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha512.Gk3 ]] 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.Gk3 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-sha256.XII 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.363 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha384.Gf1 ]] 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.Gf1 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha384.OTN 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n /tmp/spdk.key-sha256.jmj ]] 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.jmj 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@174 -- # for i in "${!keys[@]}" 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@175 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha512.AyL 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@176 -- # [[ -n '' ]] 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@179 -- # connect_authenticate sha512 ffdhe8192 3 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@65 -- # local digest dhgroup key ckey qpairs 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # digest=sha512 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # dhgroup=ffdhe8192 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@67 -- # key=key3 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@68 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@70 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@71 -- # bdev_connect -b nvme0 --dhchap-key key3 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:22.364 08:58:12 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:23.313 nvme0n1 00:22:23.313 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # hostrpc bdev_nvme_get_controllers 00:22:23.313 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # jq -r '.[].name' 00:22:23.313 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:23.313 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@73 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:23.313 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:22:23.313 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:23.313 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:23.313 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:23.313 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@74 -- # qpairs='[ 00:22:23.313 { 00:22:23.313 "cntlid": 1, 00:22:23.313 "qid": 0, 00:22:23.313 "state": "enabled", 00:22:23.313 "thread": "nvmf_tgt_poll_group_000", 00:22:23.313 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:23.313 "listen_address": { 00:22:23.313 "trtype": "TCP", 00:22:23.313 "adrfam": "IPv4", 00:22:23.313 "traddr": "10.0.0.2", 00:22:23.313 "trsvcid": "4420" 00:22:23.313 }, 00:22:23.313 "peer_address": { 00:22:23.313 "trtype": "TCP", 00:22:23.313 "adrfam": "IPv4", 00:22:23.313 "traddr": "10.0.0.1", 00:22:23.313 "trsvcid": "45288" 00:22:23.313 }, 00:22:23.313 "auth": { 00:22:23.313 "state": "completed", 00:22:23.313 "digest": "sha512", 00:22:23.313 "dhgroup": "ffdhe8192" 00:22:23.313 } 00:22:23.313 } 00:22:23.313 ]' 00:22:23.313 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # jq -r '.[0].auth.digest' 00:22:23.573 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@75 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:22:23.573 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # jq -r '.[0].auth.dhgroup' 00:22:23.573 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@76 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:22:23.573 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # jq -r '.[0].auth.state' 00:22:23.573 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@77 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:22:23.573 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@78 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:23.573 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:23.834 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@80 -- # nvme_connect --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:22:23.834 08:58:13 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@82 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:24.405 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@83 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@182 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key3 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@183 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 00:22:24.405 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 00:22:24.665 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@184 -- # NOT bdev_connect -b nvme0 --dhchap-key key3 00:22:24.665 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:22:24.665 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key3 00:22:24.665 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:22:24.665 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:24.665 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:22:24.665 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:24.665 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key3 00:22:24.665 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:24.665 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:24.927 request: 00:22:24.927 { 00:22:24.927 "name": "nvme0", 00:22:24.927 "trtype": "tcp", 00:22:24.927 "traddr": "10.0.0.2", 00:22:24.927 "adrfam": "ipv4", 00:22:24.927 "trsvcid": "4420", 00:22:24.927 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:22:24.927 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:24.927 "prchk_reftag": false, 00:22:24.927 "prchk_guard": false, 00:22:24.927 "hdgst": false, 00:22:24.927 "ddgst": false, 00:22:24.927 "dhchap_key": "key3", 00:22:24.927 "allow_unrecognized_csi": false, 00:22:24.927 "method": "bdev_nvme_attach_controller", 00:22:24.927 "req_id": 1 00:22:24.927 } 00:22:24.927 Got JSON-RPC error response 00:22:24.927 response: 00:22:24.927 { 00:22:24.927 "code": -5, 00:22:24.927 "message": "Input/output error" 00:22:24.927 } 00:22:24.927 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:22:24.927 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:24.927 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:24.927 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:24.927 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@187 -- # IFS=, 00:22:24.927 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@188 -- # printf %s sha256,sha384,sha512 00:22:24.927 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@187 -- # hostrpc bdev_nvme_set_options --dhchap-dhgroups ffdhe2048 --dhchap-digests sha256,sha384,sha512 00:22:24.927 08:58:14 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-dhgroups ffdhe2048 --dhchap-digests sha256,sha384,sha512 00:22:24.927 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@193 -- # NOT bdev_connect -b nvme0 --dhchap-key key3 00:22:24.927 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:22:24.927 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key3 00:22:24.927 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:22:24.927 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:24.927 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:22:24.927 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:24.927 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key3 00:22:24.927 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:24.927 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key3 00:22:25.188 request: 00:22:25.188 { 00:22:25.188 "name": "nvme0", 00:22:25.188 "trtype": "tcp", 00:22:25.188 "traddr": "10.0.0.2", 00:22:25.188 "adrfam": "ipv4", 00:22:25.188 "trsvcid": "4420", 00:22:25.188 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:22:25.188 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:25.188 "prchk_reftag": false, 00:22:25.188 "prchk_guard": false, 00:22:25.188 "hdgst": false, 00:22:25.188 "ddgst": false, 00:22:25.188 "dhchap_key": "key3", 00:22:25.188 "allow_unrecognized_csi": false, 00:22:25.188 "method": "bdev_nvme_attach_controller", 00:22:25.188 "req_id": 1 00:22:25.188 } 00:22:25.188 Got JSON-RPC error response 00:22:25.188 response: 00:22:25.188 { 00:22:25.188 "code": -5, 00:22:25.188 "message": "Input/output error" 00:22:25.188 } 00:22:25.188 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:22:25.188 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:25.188 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:25.188 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:25.188 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # IFS=, 00:22:25.188 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@198 -- # printf %s sha256,sha384,sha512 00:22:25.188 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # IFS=, 00:22:25.188 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@198 -- # printf %s null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:22:25.188 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@197 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:22:25.188 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@208 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@209 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@210 -- # NOT bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:22:25.449 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:22:25.709 request: 00:22:25.709 { 00:22:25.709 "name": "nvme0", 00:22:25.709 "trtype": "tcp", 00:22:25.709 "traddr": "10.0.0.2", 00:22:25.709 "adrfam": "ipv4", 00:22:25.709 "trsvcid": "4420", 00:22:25.709 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:22:25.709 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:25.709 "prchk_reftag": false, 00:22:25.709 "prchk_guard": false, 00:22:25.709 "hdgst": false, 00:22:25.709 "ddgst": false, 00:22:25.709 "dhchap_key": "key0", 00:22:25.709 "dhchap_ctrlr_key": "key1", 00:22:25.709 "allow_unrecognized_csi": false, 00:22:25.709 "method": "bdev_nvme_attach_controller", 00:22:25.709 "req_id": 1 00:22:25.709 } 00:22:25.709 Got JSON-RPC error response 00:22:25.709 response: 00:22:25.709 { 00:22:25.709 "code": -5, 00:22:25.709 "message": "Input/output error" 00:22:25.709 } 00:22:25.709 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:22:25.709 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:25.709 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:25.709 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:25.709 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@213 -- # bdev_connect -b nvme0 --dhchap-key key0 00:22:25.709 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 00:22:25.710 08:58:15 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 00:22:25.970 nvme0n1 00:22:25.970 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # hostrpc bdev_nvme_get_controllers 00:22:25.970 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # jq -r '.[].name' 00:22:25.970 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:26.231 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@214 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:26.231 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@215 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:26.231 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:26.492 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@218 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 00:22:26.492 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:26.492 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:26.492 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:26.492 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@219 -- # bdev_connect -b nvme0 --dhchap-key key1 00:22:26.492 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:22:26.492 08:58:16 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:22:27.433 nvme0n1 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # hostrpc bdev_nvme_get_controllers 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # jq -r '.[].name' 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@220 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@222 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key key3 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # hostrpc bdev_nvme_get_controllers 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # jq -r '.[].name' 00:22:27.433 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:27.694 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@223 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:27.694 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@225 -- # nvme_connect --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:22:27.694 08:58:17 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@36 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid 00d0226a-fbea-ec11-9bc7-a4bf019282be -l 0 --dhchap-secret DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: --dhchap-ctrl-secret DHHC-1:03:MjYxZTJlNzY5YjJlNzM2Yzg5NGIwYWMzZmVhMThkMTQ5ZDY0YzNlNjZjZmE1NmJkM2YyMjBiYjFkY2M0MDRlY485M4k=: 00:22:28.267 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@226 -- # nvme_get_ctrlr 00:22:28.267 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@41 -- # local dev 00:22:28.267 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@43 -- # for dev in /sys/devices/virtual/nvme-fabrics/ctl/nvme* 00:22:28.267 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # [[ nqn.2024-03.io.spdk:cnode0 == \n\q\n\.\2\0\2\4\-\0\3\.\i\o\.\s\p\d\k\:\c\n\o\d\e\0 ]] 00:22:28.267 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # echo nvme0 00:22:28.267 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@44 -- # break 00:22:28.267 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@226 -- # nctrlr=nvme0 00:22:28.267 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@227 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:28.267 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:28.528 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@228 -- # NOT bdev_connect -b nvme0 --dhchap-key key1 00:22:28.528 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:22:28.528 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg bdev_connect -b nvme0 --dhchap-key key1 00:22:28.528 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=bdev_connect 00:22:28.528 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:28.528 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t bdev_connect 00:22:28.528 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:28.528 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # bdev_connect -b nvme0 --dhchap-key key1 00:22:28.528 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:22:28.528 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key1 00:22:29.100 request: 00:22:29.100 { 00:22:29.100 "name": "nvme0", 00:22:29.100 "trtype": "tcp", 00:22:29.100 "traddr": "10.0.0.2", 00:22:29.100 "adrfam": "ipv4", 00:22:29.100 "trsvcid": "4420", 00:22:29.100 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:22:29.100 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be", 00:22:29.100 "prchk_reftag": false, 00:22:29.100 "prchk_guard": false, 00:22:29.100 "hdgst": false, 00:22:29.100 "ddgst": false, 00:22:29.100 "dhchap_key": "key1", 00:22:29.100 "allow_unrecognized_csi": false, 00:22:29.101 "method": "bdev_nvme_attach_controller", 00:22:29.101 "req_id": 1 00:22:29.101 } 00:22:29.101 Got JSON-RPC error response 00:22:29.101 response: 00:22:29.101 { 00:22:29.101 "code": -5, 00:22:29.101 "message": "Input/output error" 00:22:29.101 } 00:22:29.101 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:22:29.101 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:29.101 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:29.101 08:58:18 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:29.101 08:58:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@229 -- # bdev_connect -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:22:29.101 08:58:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:22:29.101 08:58:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:22:30.044 nvme0n1 00:22:30.044 08:58:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # hostrpc bdev_nvme_get_controllers 00:22:30.044 08:58:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # jq -r '.[].name' 00:22:30.044 08:58:19 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:30.044 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@230 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:30.044 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@231 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:30.044 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:30.305 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@233 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:30.305 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:30.305 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:30.305 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:30.305 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@234 -- # bdev_connect -b nvme0 00:22:30.305 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 00:22:30.305 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 00:22:30.566 nvme0n1 00:22:30.566 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # hostrpc bdev_nvme_get_controllers 00:22:30.567 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # jq -r '.[].name' 00:22:30.567 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:30.567 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@235 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:30.567 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@236 -- # hostrpc bdev_nvme_detach_controller nvme0 00:22:30.567 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@239 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key key3 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@240 -- # nvme_set_keys nvme0 DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: '' 2s 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@49 -- # local ctl key ckey dev timeout 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ctl=nvme0 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # key=DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ckey= 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # timeout=2s 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@52 -- # dev=/sys/devices/virtual/nvme-fabrics/ctl/nvme0 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # [[ -z DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: ]] 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # echo DHHC-1:01:NzA1OWFkM2IzMmNlMGMxMzcwYjhkMDFhZmVmOTY3NTGQHdIk: 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # [[ -z '' ]] 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # [[ -z 2s ]] 00:22:30.828 08:58:20 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # sleep 2s 00:22:32.743 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@241 -- # waitforblk nvme0n1 00:22:32.743 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1235 -- # local i=0 00:22:32.743 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1236 -- # lsblk -l -o NAME 00:22:32.743 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1236 -- # grep -q -w nvme0n1 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1242 -- # lsblk -l -o NAME 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1242 -- # grep -q -w nvme0n1 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # return 0 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@243 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key1 --dhchap-ctrlr-key key2 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@244 -- # nvme_set_keys nvme0 '' DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: 2s 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@49 -- # local ctl key ckey dev timeout 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ctl=nvme0 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # key= 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # ckey=DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@51 -- # timeout=2s 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@52 -- # dev=/sys/devices/virtual/nvme-fabrics/ctl/nvme0 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@54 -- # [[ -z '' ]] 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # [[ -z DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: ]] 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@55 -- # echo DHHC-1:02:MTMxZmZjMWRjNzczMWRlYzYyZGEwOTgyYjIxOTQwMGRiZDY2NGE0YmE0NmU3MmM1iyCweA==: 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # [[ -z 2s ]] 00:22:33.005 08:58:22 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@56 -- # sleep 2s 00:22:34.920 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@245 -- # waitforblk nvme0n1 00:22:34.920 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1235 -- # local i=0 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1236 -- # lsblk -l -o NAME 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1236 -- # grep -q -w nvme0n1 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1242 -- # lsblk -l -o NAME 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1242 -- # grep -q -w nvme0n1 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1246 -- # return 0 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@246 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:22:34.921 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@249 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key key1 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@250 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:22:34.921 08:58:24 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:22:35.862 nvme0n1 00:22:35.862 08:58:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@252 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key key3 00:22:35.862 08:58:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:35.862 08:58:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:35.862 08:58:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:35.862 08:58:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@253 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:22:35.862 08:58:25 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key3 00:22:36.433 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # hostrpc bdev_nvme_get_controllers 00:22:36.433 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # jq -r '.[].name' 00:22:36.433 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:36.694 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@254 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:36.694 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@256 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:36.694 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:36.694 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:36.694 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:36.694 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@257 -- # hostrpc bdev_nvme_set_keys nvme0 00:22:36.694 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 00:22:36.694 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # hostrpc bdev_nvme_get_controllers 00:22:36.694 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # jq -r '.[].name' 00:22:36.695 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@258 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@260 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key key3 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@261 -- # NOT hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=hostrpc 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t hostrpc 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:22:36.956 08:58:26 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key key3 00:22:37.528 request: 00:22:37.528 { 00:22:37.528 "name": "nvme0", 00:22:37.528 "dhchap_key": "key1", 00:22:37.528 "dhchap_ctrlr_key": "key3", 00:22:37.528 "method": "bdev_nvme_set_keys", 00:22:37.528 "req_id": 1 00:22:37.528 } 00:22:37.528 Got JSON-RPC error response 00:22:37.528 response: 00:22:37.528 { 00:22:37.528 "code": -13, 00:22:37.528 "message": "Permission denied" 00:22:37.528 } 00:22:37.528 08:58:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:22:37.528 08:58:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:37.528 08:58:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:37.528 08:58:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:37.528 08:58:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # hostrpc bdev_nvme_get_controllers 00:22:37.528 08:58:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # jq length 00:22:37.528 08:58:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:37.528 08:58:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # (( 1 != 0 )) 00:22:37.528 08:58:27 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@263 -- # sleep 1s 00:22:38.919 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # hostrpc bdev_nvme_get_controllers 00:22:38.919 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # jq length 00:22:38.919 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:38.919 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@262 -- # (( 0 != 0 )) 00:22:38.919 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@267 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key0 --dhchap-ctrlr-key key1 00:22:38.919 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:38.920 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:38.920 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:38.920 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@268 -- # bdev_connect -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:22:38.920 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@60 -- # hostrpc bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:22:38.920 08:58:28 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be -n nqn.2024-03.io.spdk:cnode0 -b nvme0 --dhchap-key key0 --dhchap-ctrlr-key key1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:22:39.863 nvme0n1 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@270 -- # rpc_cmd nvmf_subsystem_set_keys nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --dhchap-key key2 --dhchap-ctrlr-key key3 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@271 -- # NOT hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@650 -- # local es=0 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@652 -- # valid_exec_arg hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@638 -- # local arg=hostrpc 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # type -t hostrpc 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # hostrpc bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:22:39.863 08:58:29 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key key0 00:22:40.124 request: 00:22:40.124 { 00:22:40.124 "name": "nvme0", 00:22:40.124 "dhchap_key": "key2", 00:22:40.124 "dhchap_ctrlr_key": "key0", 00:22:40.124 "method": "bdev_nvme_set_keys", 00:22:40.124 "req_id": 1 00:22:40.124 } 00:22:40.124 Got JSON-RPC error response 00:22:40.124 response: 00:22:40.124 { 00:22:40.124 "code": -13, 00:22:40.124 "message": "Permission denied" 00:22:40.124 } 00:22:40.124 08:58:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@653 -- # es=1 00:22:40.124 08:58:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:22:40.124 08:58:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:22:40.124 08:58:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:22:40.124 08:58:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # hostrpc bdev_nvme_get_controllers 00:22:40.124 08:58:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:40.124 08:58:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # jq length 00:22:40.384 08:58:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # (( 1 != 0 )) 00:22:40.384 08:58:30 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@273 -- # sleep 1s 00:22:41.326 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # hostrpc bdev_nvme_get_controllers 00:22:41.326 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # jq length 00:22:41.326 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@272 -- # (( 0 != 0 )) 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@276 -- # trap - SIGINT SIGTERM EXIT 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@277 -- # cleanup 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@21 -- # killprocess 1866679 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@950 -- # '[' -z 1866679 ']' 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # kill -0 1866679 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # uname 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1866679 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1866679' 00:22:41.588 killing process with pid 1866679 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@969 -- # kill 1866679 00:22:41.588 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@974 -- # wait 1866679 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@22 -- # nvmftestfini 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@514 -- # nvmfcleanup 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@121 -- # sync 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@124 -- # set +e 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:22:41.849 rmmod nvme_tcp 00:22:41.849 rmmod nvme_fabrics 00:22:41.849 rmmod nvme_keyring 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@128 -- # set -e 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@129 -- # return 0 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@515 -- # '[' -n 1894314 ']' 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@516 -- # killprocess 1894314 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@950 -- # '[' -z 1894314 ']' 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@954 -- # kill -0 1894314 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # uname 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1894314 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1894314' 00:22:41.849 killing process with pid 1894314 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@969 -- # kill 1894314 00:22:41.849 08:58:31 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@974 -- # wait 1894314 00:22:42.110 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:22:42.110 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:22:42.110 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:22:42.110 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@297 -- # iptr 00:22:42.110 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@789 -- # iptables-save 00:22:42.110 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:22:42.110 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@789 -- # iptables-restore 00:22:42.111 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:22:42.111 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:22:42.111 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:42.111 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:42.111 08:58:32 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:44.025 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:22:44.025 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- target/auth.sh@23 -- # rm -f /tmp/spdk.key-null.2c1 /tmp/spdk.key-sha256.XII /tmp/spdk.key-sha384.OTN /tmp/spdk.key-sha512.AyL /tmp/spdk.key-sha512.Gk3 /tmp/spdk.key-sha384.Gf1 /tmp/spdk.key-sha256.jmj '' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf-auth.log 00:22:44.025 00:22:44.025 real 2m44.120s 00:22:44.025 user 6m7.304s 00:22:44.025 sys 0m24.130s 00:22:44.025 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:44.025 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:22:44.025 ************************************ 00:22:44.025 END TEST nvmf_auth_target 00:22:44.025 ************************************ 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@39 -- # '[' tcp = tcp ']' 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@40 -- # run_test nvmf_bdevio_no_huge /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:22:44.287 ************************************ 00:22:44.287 START TEST nvmf_bdevio_no_huge 00:22:44.287 ************************************ 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:22:44.287 * Looking for test storage... 00:22:44.287 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1689 -- # lcov --version 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@333 -- # local ver1 ver1_l 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@334 -- # local ver2 ver2_l 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@336 -- # IFS=.-: 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@336 -- # read -ra ver1 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@337 -- # IFS=.-: 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@337 -- # read -ra ver2 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@338 -- # local 'op=<' 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@340 -- # ver1_l=2 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@341 -- # ver2_l=1 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@344 -- # case "$op" in 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@345 -- # : 1 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@364 -- # (( v = 0 )) 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@365 -- # decimal 1 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@353 -- # local d=1 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@355 -- # echo 1 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@365 -- # ver1[v]=1 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@366 -- # decimal 2 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@353 -- # local d=2 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@355 -- # echo 2 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@366 -- # ver2[v]=2 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@368 -- # return 0 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:22:44.287 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:44.287 --rc genhtml_branch_coverage=1 00:22:44.287 --rc genhtml_function_coverage=1 00:22:44.287 --rc genhtml_legend=1 00:22:44.287 --rc geninfo_all_blocks=1 00:22:44.287 --rc geninfo_unexecuted_blocks=1 00:22:44.287 00:22:44.287 ' 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:22:44.287 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:44.287 --rc genhtml_branch_coverage=1 00:22:44.287 --rc genhtml_function_coverage=1 00:22:44.287 --rc genhtml_legend=1 00:22:44.287 --rc geninfo_all_blocks=1 00:22:44.287 --rc geninfo_unexecuted_blocks=1 00:22:44.287 00:22:44.287 ' 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:22:44.287 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:44.287 --rc genhtml_branch_coverage=1 00:22:44.287 --rc genhtml_function_coverage=1 00:22:44.287 --rc genhtml_legend=1 00:22:44.287 --rc geninfo_all_blocks=1 00:22:44.287 --rc geninfo_unexecuted_blocks=1 00:22:44.287 00:22:44.287 ' 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:22:44.287 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:44.287 --rc genhtml_branch_coverage=1 00:22:44.287 --rc genhtml_function_coverage=1 00:22:44.287 --rc genhtml_legend=1 00:22:44.287 --rc geninfo_all_blocks=1 00:22:44.287 --rc geninfo_unexecuted_blocks=1 00:22:44.287 00:22:44.287 ' 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:44.287 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@7 -- # uname -s 00:22:44.549 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:44.549 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:44.549 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:44.549 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:44.549 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:44.549 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@15 -- # shopt -s extglob 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@5 -- # export PATH 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@51 -- # : 0 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:22:44.550 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@55 -- # have_pci_nics=0 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@14 -- # nvmftestinit 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@474 -- # prepare_net_devs 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@436 -- # local -g is_hw=no 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@438 -- # remove_spdk_ns 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@309 -- # xtrace_disable 00:22:44.550 08:58:34 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@315 -- # pci_devs=() 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@315 -- # local -a pci_devs 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@316 -- # pci_net_devs=() 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@317 -- # pci_drivers=() 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@317 -- # local -A pci_drivers 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@319 -- # net_devs=() 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@319 -- # local -ga net_devs 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@320 -- # e810=() 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@320 -- # local -ga e810 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@321 -- # x722=() 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@321 -- # local -ga x722 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@322 -- # mlx=() 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@322 -- # local -ga mlx 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:22:52.697 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:22:52.697 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:22:52.697 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:22:52.698 Found net devices under 0000:4b:00.0: cvl_0_0 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@416 -- # [[ up == up ]] 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:22:52.698 Found net devices under 0000:4b:00.1: cvl_0_1 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@440 -- # is_hw=yes 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:22:52.698 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:52.698 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.515 ms 00:22:52.698 00:22:52.698 --- 10.0.0.2 ping statistics --- 00:22:52.698 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:52.698 rtt min/avg/max/mdev = 0.515/0.515/0.515/0.000 ms 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:52.698 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:52.698 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.331 ms 00:22:52.698 00:22:52.698 --- 10.0.0.1 ping statistics --- 00:22:52.698 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:52.698 rtt min/avg/max/mdev = 0.331/0.331/0.331/0.000 ms 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@448 -- # return 0 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@724 -- # xtrace_disable 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@507 -- # nvmfpid=1902499 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@508 -- # waitforlisten 1902499 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --no-huge -s 1024 -m 0x78 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@831 -- # '[' -z 1902499 ']' 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@836 -- # local max_retries=100 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:52.698 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@840 -- # xtrace_disable 00:22:52.698 08:58:41 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:52.698 [2024-11-06 08:58:41.874195] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:22:52.698 [2024-11-06 08:58:41.874271] [ DPDK EAL parameters: nvmf -c 0x78 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk0 --proc-type=auto ] 00:22:52.698 [2024-11-06 08:58:41.982591] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:22:52.698 [2024-11-06 08:58:42.043102] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:52.698 [2024-11-06 08:58:42.043150] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:52.698 [2024-11-06 08:58:42.043159] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:52.698 [2024-11-06 08:58:42.043166] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:52.698 [2024-11-06 08:58:42.043172] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:52.698 [2024-11-06 08:58:42.044723] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:22:52.698 [2024-11-06 08:58:42.044887] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:22:52.698 [2024-11-06 08:58:42.045144] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:22:52.698 [2024-11-06 08:58:42.045147] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@864 -- # return 0 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@730 -- # xtrace_disable 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:52.698 [2024-11-06 08:58:42.740517] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.698 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:52.698 Malloc0 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:52.699 [2024-11-06 08:58:42.778308] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 --no-huge -s 1024 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@558 -- # config=() 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@558 -- # local subsystem config 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:22:52.699 { 00:22:52.699 "params": { 00:22:52.699 "name": "Nvme$subsystem", 00:22:52.699 "trtype": "$TEST_TRANSPORT", 00:22:52.699 "traddr": "$NVMF_FIRST_TARGET_IP", 00:22:52.699 "adrfam": "ipv4", 00:22:52.699 "trsvcid": "$NVMF_PORT", 00:22:52.699 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:22:52.699 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:22:52.699 "hdgst": ${hdgst:-false}, 00:22:52.699 "ddgst": ${ddgst:-false} 00:22:52.699 }, 00:22:52.699 "method": "bdev_nvme_attach_controller" 00:22:52.699 } 00:22:52.699 EOF 00:22:52.699 )") 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@580 -- # cat 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@582 -- # jq . 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@583 -- # IFS=, 00:22:52.699 08:58:42 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:22:52.699 "params": { 00:22:52.699 "name": "Nvme1", 00:22:52.699 "trtype": "tcp", 00:22:52.699 "traddr": "10.0.0.2", 00:22:52.699 "adrfam": "ipv4", 00:22:52.699 "trsvcid": "4420", 00:22:52.699 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:22:52.699 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:22:52.699 "hdgst": false, 00:22:52.699 "ddgst": false 00:22:52.699 }, 00:22:52.699 "method": "bdev_nvme_attach_controller" 00:22:52.699 }' 00:22:52.960 [2024-11-06 08:58:42.846936] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:22:52.960 [2024-11-06 08:58:42.847006] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk_pid1902843 ] 00:22:52.960 [2024-11-06 08:58:42.928845] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:22:52.960 [2024-11-06 08:58:42.984250] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:22:52.960 [2024-11-06 08:58:42.984374] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:22:52.960 [2024-11-06 08:58:42.984378] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:22:53.226 I/O targets: 00:22:53.226 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:22:53.226 00:22:53.226 00:22:53.226 CUnit - A unit testing framework for C - Version 2.1-3 00:22:53.226 http://cunit.sourceforge.net/ 00:22:53.226 00:22:53.226 00:22:53.226 Suite: bdevio tests on: Nvme1n1 00:22:53.516 Test: blockdev write read block ...passed 00:22:53.516 Test: blockdev write zeroes read block ...passed 00:22:53.516 Test: blockdev write zeroes read no split ...passed 00:22:53.516 Test: blockdev write zeroes read split ...passed 00:22:53.516 Test: blockdev write zeroes read split partial ...passed 00:22:53.516 Test: blockdev reset ...[2024-11-06 08:58:43.496090] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:22:53.516 [2024-11-06 08:58:43.496160] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x217a800 (9): Bad file descriptor 00:22:53.517 [2024-11-06 08:58:43.553656] bdev_nvme.c:2236:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller successful. 00:22:53.517 passed 00:22:53.517 Test: blockdev write read 8 blocks ...passed 00:22:53.517 Test: blockdev write read size > 128k ...passed 00:22:53.517 Test: blockdev write read invalid size ...passed 00:22:53.823 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:22:53.823 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:22:53.823 Test: blockdev write read max offset ...passed 00:22:53.823 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:22:53.823 Test: blockdev writev readv 8 blocks ...passed 00:22:53.823 Test: blockdev writev readv 30 x 1block ...passed 00:22:53.823 Test: blockdev writev readv block ...passed 00:22:53.823 Test: blockdev writev readv size > 128k ...passed 00:22:53.823 Test: blockdev writev readv size > 128k in two iovs ...passed 00:22:53.823 Test: blockdev comparev and writev ...[2024-11-06 08:58:43.778293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:22:53.823 [2024-11-06 08:58:43.778320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:22:53.823 [2024-11-06 08:58:43.778331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:22:53.823 [2024-11-06 08:58:43.778337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:22:53.823 [2024-11-06 08:58:43.778848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:22:53.823 [2024-11-06 08:58:43.778857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:22:53.823 [2024-11-06 08:58:43.778868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:22:53.823 [2024-11-06 08:58:43.778875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:22:53.823 [2024-11-06 08:58:43.779366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:22:53.823 [2024-11-06 08:58:43.779374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:22:53.823 [2024-11-06 08:58:43.779384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:22:53.823 [2024-11-06 08:58:43.779389] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:22:53.823 [2024-11-06 08:58:43.779836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:22:53.823 [2024-11-06 08:58:43.779846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:22:53.823 [2024-11-06 08:58:43.779855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:22:53.823 [2024-11-06 08:58:43.779861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:22:53.823 passed 00:22:53.823 Test: blockdev nvme passthru rw ...passed 00:22:53.823 Test: blockdev nvme passthru vendor specific ...[2024-11-06 08:58:43.864550] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:22:53.823 [2024-11-06 08:58:43.864561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:22:53.823 [2024-11-06 08:58:43.864880] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:22:53.823 [2024-11-06 08:58:43.864888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:22:53.823 [2024-11-06 08:58:43.865246] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:22:53.823 [2024-11-06 08:58:43.865254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:22:53.823 [2024-11-06 08:58:43.865563] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:22:53.823 [2024-11-06 08:58:43.865572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:22:53.823 passed 00:22:53.823 Test: blockdev nvme admin passthru ...passed 00:22:54.100 Test: blockdev copy ...passed 00:22:54.100 00:22:54.100 Run Summary: Type Total Ran Passed Failed Inactive 00:22:54.100 suites 1 1 n/a 0 0 00:22:54.100 tests 23 23 23 0 0 00:22:54.100 asserts 152 152 152 0 n/a 00:22:54.100 00:22:54.100 Elapsed time = 1.220 seconds 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@561 -- # xtrace_disable 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- target/bdevio.sh@30 -- # nvmftestfini 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@514 -- # nvmfcleanup 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@121 -- # sync 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@124 -- # set +e 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@125 -- # for i in {1..20} 00:22:54.100 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:22:54.100 rmmod nvme_tcp 00:22:54.362 rmmod nvme_fabrics 00:22:54.362 rmmod nvme_keyring 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@128 -- # set -e 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@129 -- # return 0 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@515 -- # '[' -n 1902499 ']' 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@516 -- # killprocess 1902499 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@950 -- # '[' -z 1902499 ']' 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@954 -- # kill -0 1902499 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@955 -- # uname 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1902499 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@956 -- # process_name=reactor_3 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@960 -- # '[' reactor_3 = sudo ']' 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1902499' 00:22:54.362 killing process with pid 1902499 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@969 -- # kill 1902499 00:22:54.362 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@974 -- # wait 1902499 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@297 -- # iptr 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@789 -- # iptables-save 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@789 -- # iptables-restore 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@302 -- # remove_spdk_ns 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:54.624 08:58:44 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:22:57.173 00:22:57.173 real 0m12.595s 00:22:57.173 user 0m14.967s 00:22:57.173 sys 0m6.626s 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@1126 -- # xtrace_disable 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:22:57.173 ************************************ 00:22:57.173 END TEST nvmf_bdevio_no_huge 00:22:57.173 ************************************ 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@41 -- # run_test nvmf_tls /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:22:57.173 ************************************ 00:22:57.173 START TEST nvmf_tls 00:22:57.173 ************************************ 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:22:57.173 * Looking for test storage... 00:22:57.173 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1689 -- # lcov --version 00:22:57.173 08:58:46 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@333 -- # local ver1 ver1_l 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@334 -- # local ver2 ver2_l 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@336 -- # IFS=.-: 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@336 -- # read -ra ver1 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@337 -- # IFS=.-: 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@337 -- # read -ra ver2 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@338 -- # local 'op=<' 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@340 -- # ver1_l=2 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@341 -- # ver2_l=1 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@344 -- # case "$op" in 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@345 -- # : 1 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@364 -- # (( v = 0 )) 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@365 -- # decimal 1 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@353 -- # local d=1 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@355 -- # echo 1 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@365 -- # ver1[v]=1 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@366 -- # decimal 2 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@353 -- # local d=2 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@355 -- # echo 2 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@366 -- # ver2[v]=2 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@368 -- # return 0 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:22:57.173 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:57.173 --rc genhtml_branch_coverage=1 00:22:57.173 --rc genhtml_function_coverage=1 00:22:57.173 --rc genhtml_legend=1 00:22:57.173 --rc geninfo_all_blocks=1 00:22:57.173 --rc geninfo_unexecuted_blocks=1 00:22:57.173 00:22:57.173 ' 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:22:57.173 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:57.173 --rc genhtml_branch_coverage=1 00:22:57.173 --rc genhtml_function_coverage=1 00:22:57.173 --rc genhtml_legend=1 00:22:57.173 --rc geninfo_all_blocks=1 00:22:57.173 --rc geninfo_unexecuted_blocks=1 00:22:57.173 00:22:57.173 ' 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:22:57.173 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:57.173 --rc genhtml_branch_coverage=1 00:22:57.173 --rc genhtml_function_coverage=1 00:22:57.173 --rc genhtml_legend=1 00:22:57.173 --rc geninfo_all_blocks=1 00:22:57.173 --rc geninfo_unexecuted_blocks=1 00:22:57.173 00:22:57.173 ' 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:22:57.173 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:22:57.173 --rc genhtml_branch_coverage=1 00:22:57.173 --rc genhtml_function_coverage=1 00:22:57.173 --rc genhtml_legend=1 00:22:57.173 --rc geninfo_all_blocks=1 00:22:57.173 --rc geninfo_unexecuted_blocks=1 00:22:57.173 00:22:57.173 ' 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@7 -- # uname -s 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:57.173 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@15 -- # shopt -s extglob 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@5 -- # export PATH 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@51 -- # : 0 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:22:57.174 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@55 -- # have_pci_nics=0 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@63 -- # nvmftestinit 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@474 -- # prepare_net_devs 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@436 -- # local -g is_hw=no 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@438 -- # remove_spdk_ns 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@309 -- # xtrace_disable 00:22:57.174 08:58:47 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@315 -- # pci_devs=() 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@315 -- # local -a pci_devs 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@316 -- # pci_net_devs=() 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@317 -- # pci_drivers=() 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@317 -- # local -A pci_drivers 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@319 -- # net_devs=() 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@319 -- # local -ga net_devs 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@320 -- # e810=() 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@320 -- # local -ga e810 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@321 -- # x722=() 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@321 -- # local -ga x722 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@322 -- # mlx=() 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@322 -- # local -ga mlx 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:23:05.323 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:23:05.323 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@416 -- # [[ up == up ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:23:05.323 Found net devices under 0000:4b:00.0: cvl_0_0 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@416 -- # [[ up == up ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:23:05.323 Found net devices under 0000:4b:00.1: cvl_0_1 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@440 -- # is_hw=yes 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:05.323 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:23:05.324 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:05.324 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.718 ms 00:23:05.324 00:23:05.324 --- 10.0.0.2 ping statistics --- 00:23:05.324 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:05.324 rtt min/avg/max/mdev = 0.718/0.718/0.718/0.000 ms 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:05.324 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:05.324 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.330 ms 00:23:05.324 00:23:05.324 --- 10.0.0.1 ping statistics --- 00:23:05.324 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:05.324 rtt min/avg/max/mdev = 0.330/0.330/0.330/0.000 ms 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@448 -- # return 0 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@64 -- # nvmfappstart -m 0x2 --wait-for-rpc 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=1907307 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 1907307 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 --wait-for-rpc 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1907307 ']' 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:05.324 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:05.324 08:58:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:05.324 [2024-11-06 08:58:54.561939] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:05.324 [2024-11-06 08:58:54.562012] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:05.324 [2024-11-06 08:58:54.665034] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:05.324 [2024-11-06 08:58:54.715700] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:05.324 [2024-11-06 08:58:54.715764] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:05.324 [2024-11-06 08:58:54.715773] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:05.324 [2024-11-06 08:58:54.715780] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:05.324 [2024-11-06 08:58:54.715794] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:05.324 [2024-11-06 08:58:54.716619] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:05.324 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:05.324 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:05.324 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:23:05.324 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:23:05.324 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:05.324 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:05.324 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@66 -- # '[' tcp '!=' tcp ']' 00:23:05.324 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_set_default_impl -i ssl 00:23:05.586 true 00:23:05.586 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:23:05.586 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # jq -r .tls_version 00:23:05.847 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@74 -- # version=0 00:23:05.847 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@75 -- # [[ 0 != \0 ]] 00:23:05.847 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@81 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:23:06.109 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:23:06.109 08:58:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # jq -r .tls_version 00:23:06.109 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@82 -- # version=13 00:23:06.109 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@83 -- # [[ 13 != \1\3 ]] 00:23:06.109 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 7 00:23:06.371 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:23:06.371 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # jq -r .tls_version 00:23:06.633 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@90 -- # version=7 00:23:06.633 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@91 -- # [[ 7 != \7 ]] 00:23:06.633 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:23:06.633 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # jq -r .enable_ktls 00:23:06.633 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@97 -- # ktls=false 00:23:06.633 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@98 -- # [[ false != \f\a\l\s\e ]] 00:23:06.633 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@104 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --enable-ktls 00:23:06.894 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:23:06.894 08:58:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # jq -r .enable_ktls 00:23:07.156 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@105 -- # ktls=true 00:23:07.156 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@106 -- # [[ true != \t\r\u\e ]] 00:23:07.156 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@112 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --disable-ktls 00:23:07.156 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:23:07.156 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # jq -r .enable_ktls 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@113 -- # ktls=false 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@114 -- # [[ false != \f\a\l\s\e ]] 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@119 -- # format_interchange_psk 00112233445566778899aabbccddeeff 1 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 1 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@728 -- # local prefix key digest 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # key=00112233445566778899aabbccddeeff 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # digest=1 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@731 -- # python - 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@119 -- # key=NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@120 -- # format_interchange_psk ffeeddccbbaa99887766554433221100 1 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 ffeeddccbbaa99887766554433221100 1 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@728 -- # local prefix key digest 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # key=ffeeddccbbaa99887766554433221100 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # digest=1 00:23:07.418 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@731 -- # python - 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@120 -- # key_2=NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@122 -- # mktemp 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@122 -- # key_path=/tmp/tmp.1JPhleGAgO 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@123 -- # mktemp 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@123 -- # key_2_path=/tmp/tmp.8qvsRXAtTA 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@125 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@126 -- # echo -n NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@128 -- # chmod 0600 /tmp/tmp.1JPhleGAgO 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@129 -- # chmod 0600 /tmp/tmp.8qvsRXAtTA 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:23:07.680 08:58:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@132 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_start_init 00:23:07.942 08:58:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@134 -- # setup_nvmf_tgt /tmp/tmp.1JPhleGAgO 00:23:07.942 08:58:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.1JPhleGAgO 00:23:07.942 08:58:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:23:08.202 [2024-11-06 08:58:58.187122] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:08.202 08:58:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:23:08.464 08:58:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:23:08.464 [2024-11-06 08:58:58.536008] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:23:08.464 [2024-11-06 08:58:58.536355] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:08.464 08:58:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:23:08.725 malloc0 00:23:08.725 08:58:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:23:08.986 08:58:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.1JPhleGAgO 00:23:08.986 08:58:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:23:09.246 08:58:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@138 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -S ssl -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 hostnqn:nqn.2016-06.io.spdk:host1' --psk-path /tmp/tmp.1JPhleGAgO 00:23:19.255 Initializing NVMe Controllers 00:23:19.255 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:23:19.255 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:23:19.255 Initialization complete. Launching workers. 00:23:19.255 ======================================================== 00:23:19.255 Latency(us) 00:23:19.255 Device Information : IOPS MiB/s Average min max 00:23:19.255 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 18554.39 72.48 3449.38 872.73 4254.54 00:23:19.255 ======================================================== 00:23:19.255 Total : 18554.39 72.48 3449.38 872.73 4254.54 00:23:19.255 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@144 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.1JPhleGAgO 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.1JPhleGAgO 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=1910254 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 1910254 /var/tmp/bdevperf.sock 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1910254 ']' 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:19.255 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:19.255 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:23:19.516 [2024-11-06 08:59:09.395413] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:19.516 [2024-11-06 08:59:09.395471] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1910254 ] 00:23:19.516 [2024-11-06 08:59:09.453494] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:19.516 [2024-11-06 08:59:09.482753] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:19.516 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:19.516 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:19.516 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.1JPhleGAgO 00:23:19.775 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:23:20.035 [2024-11-06 08:59:09.888974] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:23:20.035 TLSTESTn1 00:23:20.035 08:59:09 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:23:20.035 Running I/O for 10 seconds... 00:23:22.361 6417.00 IOPS, 25.07 MiB/s [2024-11-06T07:59:13.416Z] 5958.00 IOPS, 23.27 MiB/s [2024-11-06T07:59:14.358Z] 5997.00 IOPS, 23.43 MiB/s [2024-11-06T07:59:15.301Z] 5947.00 IOPS, 23.23 MiB/s [2024-11-06T07:59:16.242Z] 6060.00 IOPS, 23.67 MiB/s [2024-11-06T07:59:17.183Z] 6146.83 IOPS, 24.01 MiB/s [2024-11-06T07:59:18.124Z] 6054.86 IOPS, 23.65 MiB/s [2024-11-06T07:59:19.507Z] 5970.38 IOPS, 23.32 MiB/s [2024-11-06T07:59:20.448Z] 5919.78 IOPS, 23.12 MiB/s [2024-11-06T07:59:20.448Z] 5981.90 IOPS, 23.37 MiB/s 00:23:30.335 Latency(us) 00:23:30.335 [2024-11-06T07:59:20.448Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:30.335 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:23:30.335 Verification LBA range: start 0x0 length 0x2000 00:23:30.335 TLSTESTn1 : 10.01 5987.42 23.39 0.00 0.00 21347.98 4969.81 33204.91 00:23:30.335 [2024-11-06T07:59:20.448Z] =================================================================================================================== 00:23:30.335 [2024-11-06T07:59:20.448Z] Total : 5987.42 23.39 0.00 0.00 21347.98 4969.81 33204.91 00:23:30.335 { 00:23:30.335 "results": [ 00:23:30.335 { 00:23:30.335 "job": "TLSTESTn1", 00:23:30.335 "core_mask": "0x4", 00:23:30.335 "workload": "verify", 00:23:30.335 "status": "finished", 00:23:30.335 "verify_range": { 00:23:30.335 "start": 0, 00:23:30.335 "length": 8192 00:23:30.335 }, 00:23:30.335 "queue_depth": 128, 00:23:30.335 "io_size": 4096, 00:23:30.335 "runtime": 10.011654, 00:23:30.335 "iops": 5987.42225810041, 00:23:30.335 "mibps": 23.388368195704725, 00:23:30.335 "io_failed": 0, 00:23:30.335 "io_timeout": 0, 00:23:30.335 "avg_latency_us": 21347.97905244895, 00:23:30.335 "min_latency_us": 4969.8133333333335, 00:23:30.335 "max_latency_us": 33204.90666666667 00:23:30.335 } 00:23:30.335 ], 00:23:30.335 "core_count": 1 00:23:30.335 } 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@45 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@46 -- # killprocess 1910254 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1910254 ']' 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1910254 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1910254 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1910254' 00:23:30.335 killing process with pid 1910254 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1910254 00:23:30.335 Received shutdown signal, test time was about 10.000000 seconds 00:23:30.335 00:23:30.335 Latency(us) 00:23:30.335 [2024-11-06T07:59:20.448Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:30.335 [2024-11-06T07:59:20.448Z] =================================================================================================================== 00:23:30.335 [2024-11-06T07:59:20.448Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1910254 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@147 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.8qvsRXAtTA 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.8qvsRXAtTA 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=run_bdevperf 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t run_bdevperf 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.8qvsRXAtTA 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.8qvsRXAtTA 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=1912288 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 1912288 /var/tmp/bdevperf.sock 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1912288 ']' 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:30.335 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:30.335 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:30.335 [2024-11-06 08:59:20.347727] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:30.335 [2024-11-06 08:59:20.347799] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1912288 ] 00:23:30.335 [2024-11-06 08:59:20.406284] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:30.335 [2024-11-06 08:59:20.435150] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:30.595 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:30.595 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:30.595 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.8qvsRXAtTA 00:23:30.595 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:23:30.856 [2024-11-06 08:59:20.845485] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:23:30.856 [2024-11-06 08:59:20.855984] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:23:30.856 [2024-11-06 08:59:20.856686] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x11cbbb0 (107): Transport endpoint is not connected 00:23:30.856 [2024-11-06 08:59:20.857681] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x11cbbb0 (9): Bad file descriptor 00:23:30.856 [2024-11-06 08:59:20.858683] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] Ctrlr is in error state 00:23:30.856 [2024-11-06 08:59:20.858691] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:23:30.856 [2024-11-06 08:59:20.858697] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode1, Operation not permitted 00:23:30.856 [2024-11-06 08:59:20.858705] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] in failed state. 00:23:30.856 request: 00:23:30.856 { 00:23:30.856 "name": "TLSTEST", 00:23:30.856 "trtype": "tcp", 00:23:30.856 "traddr": "10.0.0.2", 00:23:30.856 "adrfam": "ipv4", 00:23:30.856 "trsvcid": "4420", 00:23:30.856 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:23:30.856 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:23:30.856 "prchk_reftag": false, 00:23:30.856 "prchk_guard": false, 00:23:30.856 "hdgst": false, 00:23:30.856 "ddgst": false, 00:23:30.856 "psk": "key0", 00:23:30.856 "allow_unrecognized_csi": false, 00:23:30.856 "method": "bdev_nvme_attach_controller", 00:23:30.856 "req_id": 1 00:23:30.856 } 00:23:30.856 Got JSON-RPC error response 00:23:30.856 response: 00:23:30.856 { 00:23:30.856 "code": -5, 00:23:30.856 "message": "Input/output error" 00:23:30.856 } 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 1912288 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1912288 ']' 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1912288 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1912288 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1912288' 00:23:30.856 killing process with pid 1912288 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1912288 00:23:30.856 Received shutdown signal, test time was about 10.000000 seconds 00:23:30.856 00:23:30.856 Latency(us) 00:23:30.856 [2024-11-06T07:59:20.969Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:30.856 [2024-11-06T07:59:20.969Z] =================================================================================================================== 00:23:30.856 [2024-11-06T07:59:20.969Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:23:30.856 08:59:20 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1912288 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@150 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.1JPhleGAgO 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.1JPhleGAgO 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=run_bdevperf 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t run_bdevperf 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.1JPhleGAgO 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host2 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.1JPhleGAgO 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=1912603 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 1912603 /var/tmp/bdevperf.sock 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1912603 ']' 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:31.117 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:31.117 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:31.117 [2024-11-06 08:59:21.103507] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:31.117 [2024-11-06 08:59:21.103567] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1912603 ] 00:23:31.117 [2024-11-06 08:59:21.162013] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:31.117 [2024-11-06 08:59:21.188695] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:31.378 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:31.378 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:31.378 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.1JPhleGAgO 00:23:31.378 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 --psk key0 00:23:31.638 [2024-11-06 08:59:21.602803] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:23:31.638 [2024-11-06 08:59:21.607732] tcp.c: 969:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:23:31.638 [2024-11-06 08:59:21.607757] posix.c: 574:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:23:31.638 [2024-11-06 08:59:21.607777] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:23:31.638 [2024-11-06 08:59:21.607964] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xa6abb0 (107): Transport endpoint is not connected 00:23:31.638 [2024-11-06 08:59:21.608952] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xa6abb0 (9): Bad file descriptor 00:23:31.638 [2024-11-06 08:59:21.609954] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] Ctrlr is in error state 00:23:31.638 [2024-11-06 08:59:21.609962] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:23:31.638 [2024-11-06 08:59:21.609968] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode1, Operation not permitted 00:23:31.638 [2024-11-06 08:59:21.609976] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 0] in failed state. 00:23:31.638 request: 00:23:31.638 { 00:23:31.638 "name": "TLSTEST", 00:23:31.638 "trtype": "tcp", 00:23:31.638 "traddr": "10.0.0.2", 00:23:31.638 "adrfam": "ipv4", 00:23:31.638 "trsvcid": "4420", 00:23:31.638 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:23:31.638 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:23:31.638 "prchk_reftag": false, 00:23:31.638 "prchk_guard": false, 00:23:31.638 "hdgst": false, 00:23:31.638 "ddgst": false, 00:23:31.638 "psk": "key0", 00:23:31.638 "allow_unrecognized_csi": false, 00:23:31.638 "method": "bdev_nvme_attach_controller", 00:23:31.638 "req_id": 1 00:23:31.638 } 00:23:31.638 Got JSON-RPC error response 00:23:31.638 response: 00:23:31.638 { 00:23:31.638 "code": -5, 00:23:31.638 "message": "Input/output error" 00:23:31.638 } 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 1912603 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1912603 ']' 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1912603 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1912603 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1912603' 00:23:31.638 killing process with pid 1912603 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1912603 00:23:31.638 Received shutdown signal, test time was about 10.000000 seconds 00:23:31.638 00:23:31.638 Latency(us) 00:23:31.638 [2024-11-06T07:59:21.751Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:31.638 [2024-11-06T07:59:21.751Z] =================================================================================================================== 00:23:31.638 [2024-11-06T07:59:21.751Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:23:31.638 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1912603 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@153 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.1JPhleGAgO 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.1JPhleGAgO 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=run_bdevperf 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t run_bdevperf 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.1JPhleGAgO 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode2 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.1JPhleGAgO 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=1912625 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 1912625 /var/tmp/bdevperf.sock 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1912625 ']' 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:31.900 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:31.900 08:59:21 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:31.900 [2024-11-06 08:59:21.851673] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:31.900 [2024-11-06 08:59:21.851727] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1912625 ] 00:23:31.900 [2024-11-06 08:59:21.910149] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:31.900 [2024-11-06 08:59:21.938074] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:32.161 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:32.161 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:32.161 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.1JPhleGAgO 00:23:32.161 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -q nqn.2016-06.io.spdk:host1 --psk key0 00:23:32.422 [2024-11-06 08:59:22.360264] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:23:32.422 [2024-11-06 08:59:22.369577] tcp.c: 969:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:23:32.422 [2024-11-06 08:59:22.369595] posix.c: 574:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:23:32.422 [2024-11-06 08:59:22.369614] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:23:32.422 [2024-11-06 08:59:22.370481] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1391bb0 (107): Transport endpoint is not connected 00:23:32.422 [2024-11-06 08:59:22.371477] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1391bb0 (9): Bad file descriptor 00:23:32.422 [2024-11-06 08:59:22.372479] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 0] Ctrlr is in error state 00:23:32.422 [2024-11-06 08:59:22.372488] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:23:32.422 [2024-11-06 08:59:22.372494] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode2, Operation not permitted 00:23:32.422 [2024-11-06 08:59:22.372502] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 0] in failed state. 00:23:32.422 request: 00:23:32.422 { 00:23:32.422 "name": "TLSTEST", 00:23:32.422 "trtype": "tcp", 00:23:32.422 "traddr": "10.0.0.2", 00:23:32.422 "adrfam": "ipv4", 00:23:32.422 "trsvcid": "4420", 00:23:32.422 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:23:32.422 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:23:32.422 "prchk_reftag": false, 00:23:32.422 "prchk_guard": false, 00:23:32.422 "hdgst": false, 00:23:32.422 "ddgst": false, 00:23:32.422 "psk": "key0", 00:23:32.422 "allow_unrecognized_csi": false, 00:23:32.422 "method": "bdev_nvme_attach_controller", 00:23:32.422 "req_id": 1 00:23:32.422 } 00:23:32.422 Got JSON-RPC error response 00:23:32.422 response: 00:23:32.422 { 00:23:32.422 "code": -5, 00:23:32.422 "message": "Input/output error" 00:23:32.422 } 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 1912625 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1912625 ']' 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1912625 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1912625 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1912625' 00:23:32.422 killing process with pid 1912625 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1912625 00:23:32.422 Received shutdown signal, test time was about 10.000000 seconds 00:23:32.422 00:23:32.422 Latency(us) 00:23:32.422 [2024-11-06T07:59:22.535Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:32.422 [2024-11-06T07:59:22.535Z] =================================================================================================================== 00:23:32.422 [2024-11-06T07:59:22.535Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:23:32.422 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1912625 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@156 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=run_bdevperf 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t run_bdevperf 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk= 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=1912906 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 1912906 /var/tmp/bdevperf.sock 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1912906 ']' 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:32.684 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:32.684 [2024-11-06 08:59:22.617923] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:32.684 [2024-11-06 08:59:22.617978] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1912906 ] 00:23:32.684 [2024-11-06 08:59:22.676140] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:32.684 [2024-11-06 08:59:22.704622] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:32.684 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 '' 00:23:32.944 [2024-11-06 08:59:22.942223] keyring.c: 24:keyring_file_check_path: *ERROR*: Non-absolute paths are not allowed: 00:23:32.944 [2024-11-06 08:59:22.942253] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:23:32.944 request: 00:23:32.944 { 00:23:32.944 "name": "key0", 00:23:32.944 "path": "", 00:23:32.944 "method": "keyring_file_add_key", 00:23:32.944 "req_id": 1 00:23:32.944 } 00:23:32.944 Got JSON-RPC error response 00:23:32.944 response: 00:23:32.944 { 00:23:32.944 "code": -1, 00:23:32.944 "message": "Operation not permitted" 00:23:32.944 } 00:23:32.944 08:59:22 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:23:33.204 [2024-11-06 08:59:23.122769] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:23:33.204 [2024-11-06 08:59:23.122793] bdev_nvme.c:6529:spdk_bdev_nvme_create: *ERROR*: Could not load PSK: key0 00:23:33.204 request: 00:23:33.204 { 00:23:33.204 "name": "TLSTEST", 00:23:33.204 "trtype": "tcp", 00:23:33.204 "traddr": "10.0.0.2", 00:23:33.204 "adrfam": "ipv4", 00:23:33.204 "trsvcid": "4420", 00:23:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:23:33.204 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:23:33.204 "prchk_reftag": false, 00:23:33.204 "prchk_guard": false, 00:23:33.204 "hdgst": false, 00:23:33.204 "ddgst": false, 00:23:33.204 "psk": "key0", 00:23:33.204 "allow_unrecognized_csi": false, 00:23:33.204 "method": "bdev_nvme_attach_controller", 00:23:33.204 "req_id": 1 00:23:33.204 } 00:23:33.204 Got JSON-RPC error response 00:23:33.204 response: 00:23:33.204 { 00:23:33.204 "code": -126, 00:23:33.204 "message": "Required key not available" 00:23:33.204 } 00:23:33.204 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 1912906 00:23:33.204 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1912906 ']' 00:23:33.204 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1912906 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1912906 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1912906' 00:23:33.205 killing process with pid 1912906 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1912906 00:23:33.205 Received shutdown signal, test time was about 10.000000 seconds 00:23:33.205 00:23:33.205 Latency(us) 00:23:33.205 [2024-11-06T07:59:23.318Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:33.205 [2024-11-06T07:59:23.318Z] =================================================================================================================== 00:23:33.205 [2024-11-06T07:59:23.318Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1912906 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@159 -- # killprocess 1907307 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1907307 ']' 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1907307 00:23:33.205 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1907307 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1907307' 00:23:33.466 killing process with pid 1907307 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1907307 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1907307 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@160 -- # format_interchange_psk 00112233445566778899aabbccddeeff0011223344556677 2 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff0011223344556677 2 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@728 -- # local prefix key digest 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # key=00112233445566778899aabbccddeeff0011223344556677 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@730 -- # digest=2 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@731 -- # python - 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@160 -- # key_long=NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@161 -- # mktemp 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@161 -- # key_long_path=/tmp/tmp.mY106GdROb 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@162 -- # echo -n NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@163 -- # chmod 0600 /tmp/tmp.mY106GdROb 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@164 -- # nvmfappstart -m 0x2 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=1912985 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 1912985 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1912985 ']' 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:33.466 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:33.466 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:33.726 [2024-11-06 08:59:23.606000] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:33.726 [2024-11-06 08:59:23.606054] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:33.726 [2024-11-06 08:59:23.672356] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:33.726 [2024-11-06 08:59:23.700149] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:33.726 [2024-11-06 08:59:23.700181] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:33.726 [2024-11-06 08:59:23.700187] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:33.726 [2024-11-06 08:59:23.700192] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:33.726 [2024-11-06 08:59:23.700196] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:33.726 [2024-11-06 08:59:23.700671] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:33.726 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:33.726 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:33.726 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:23:33.726 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:23:33.726 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:33.726 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:33.726 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@166 -- # setup_nvmf_tgt /tmp/tmp.mY106GdROb 00:23:33.726 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.mY106GdROb 00:23:33.726 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:23:33.987 [2024-11-06 08:59:23.967876] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:33.987 08:59:23 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:23:34.248 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:23:34.248 [2024-11-06 08:59:24.292672] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:23:34.248 [2024-11-06 08:59:24.292869] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:34.248 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:23:34.508 malloc0 00:23:34.508 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:23:34.508 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.mY106GdROb 00:23:34.770 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@168 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.mY106GdROb 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.mY106GdROb 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=1913347 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 1913347 /var/tmp/bdevperf.sock 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1913347 ']' 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:35.031 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:35.031 08:59:24 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:35.031 [2024-11-06 08:59:24.975643] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:35.031 [2024-11-06 08:59:24.975707] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1913347 ] 00:23:35.031 [2024-11-06 08:59:25.034945] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:35.031 [2024-11-06 08:59:25.063840] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:35.292 08:59:25 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:35.292 08:59:25 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:35.292 08:59:25 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.mY106GdROb 00:23:35.292 08:59:25 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:23:35.552 [2024-11-06 08:59:25.469993] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:23:35.552 TLSTESTn1 00:23:35.552 08:59:25 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:23:35.552 Running I/O for 10 seconds... 00:23:37.880 6355.00 IOPS, 24.82 MiB/s [2024-11-06T07:59:28.936Z] 6019.00 IOPS, 23.51 MiB/s [2024-11-06T07:59:29.880Z] 5979.67 IOPS, 23.36 MiB/s [2024-11-06T07:59:30.821Z] 6049.75 IOPS, 23.63 MiB/s [2024-11-06T07:59:31.764Z] 6140.00 IOPS, 23.98 MiB/s [2024-11-06T07:59:32.706Z] 6198.17 IOPS, 24.21 MiB/s [2024-11-06T07:59:34.090Z] 6194.71 IOPS, 24.20 MiB/s [2024-11-06T07:59:35.031Z] 6088.25 IOPS, 23.78 MiB/s [2024-11-06T07:59:35.972Z] 6123.22 IOPS, 23.92 MiB/s [2024-11-06T07:59:35.972Z] 6118.60 IOPS, 23.90 MiB/s 00:23:45.859 Latency(us) 00:23:45.859 [2024-11-06T07:59:35.972Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:45.859 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:23:45.859 Verification LBA range: start 0x0 length 0x2000 00:23:45.859 TLSTESTn1 : 10.05 6102.34 23.84 0.00 0.00 20911.10 6444.37 44346.03 00:23:45.859 [2024-11-06T07:59:35.972Z] =================================================================================================================== 00:23:45.859 [2024-11-06T07:59:35.972Z] Total : 6102.34 23.84 0.00 0.00 20911.10 6444.37 44346.03 00:23:45.859 { 00:23:45.859 "results": [ 00:23:45.859 { 00:23:45.859 "job": "TLSTESTn1", 00:23:45.859 "core_mask": "0x4", 00:23:45.859 "workload": "verify", 00:23:45.859 "status": "finished", 00:23:45.859 "verify_range": { 00:23:45.859 "start": 0, 00:23:45.859 "length": 8192 00:23:45.859 }, 00:23:45.859 "queue_depth": 128, 00:23:45.859 "io_size": 4096, 00:23:45.859 "runtime": 10.047629, 00:23:45.859 "iops": 6102.33518773434, 00:23:45.859 "mibps": 23.837246827087267, 00:23:45.859 "io_failed": 0, 00:23:45.859 "io_timeout": 0, 00:23:45.859 "avg_latency_us": 20911.099869306632, 00:23:45.859 "min_latency_us": 6444.373333333333, 00:23:45.859 "max_latency_us": 44346.026666666665 00:23:45.859 } 00:23:45.859 ], 00:23:45.859 "core_count": 1 00:23:45.859 } 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@45 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@46 -- # killprocess 1913347 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1913347 ']' 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1913347 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1913347 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1913347' 00:23:45.859 killing process with pid 1913347 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1913347 00:23:45.859 Received shutdown signal, test time was about 10.000000 seconds 00:23:45.859 00:23:45.859 Latency(us) 00:23:45.859 [2024-11-06T07:59:35.972Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:45.859 [2024-11-06T07:59:35.972Z] =================================================================================================================== 00:23:45.859 [2024-11-06T07:59:35.972Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1913347 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@171 -- # chmod 0666 /tmp/tmp.mY106GdROb 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@172 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.mY106GdROb 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.mY106GdROb 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=run_bdevperf 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t run_bdevperf 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.mY106GdROb 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@23 -- # psk=/tmp/tmp.mY106GdROb 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=1915388 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 1915388 /var/tmp/bdevperf.sock 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1915388 ']' 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:45.859 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:45.859 08:59:35 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:46.120 [2024-11-06 08:59:35.974352] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:46.120 [2024-11-06 08:59:35.974424] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1915388 ] 00:23:46.120 [2024-11-06 08:59:36.034545] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:46.120 [2024-11-06 08:59:36.062939] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:46.120 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:46.120 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:46.120 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.mY106GdROb 00:23:46.381 [2024-11-06 08:59:36.296953] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.mY106GdROb': 0100666 00:23:46.381 [2024-11-06 08:59:36.296982] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:23:46.381 request: 00:23:46.381 { 00:23:46.381 "name": "key0", 00:23:46.381 "path": "/tmp/tmp.mY106GdROb", 00:23:46.381 "method": "keyring_file_add_key", 00:23:46.381 "req_id": 1 00:23:46.381 } 00:23:46.381 Got JSON-RPC error response 00:23:46.381 response: 00:23:46.381 { 00:23:46.381 "code": -1, 00:23:46.381 "message": "Operation not permitted" 00:23:46.381 } 00:23:46.381 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:23:46.381 [2024-11-06 08:59:36.481494] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:23:46.381 [2024-11-06 08:59:36.481522] bdev_nvme.c:6529:spdk_bdev_nvme_create: *ERROR*: Could not load PSK: key0 00:23:46.381 request: 00:23:46.381 { 00:23:46.381 "name": "TLSTEST", 00:23:46.381 "trtype": "tcp", 00:23:46.381 "traddr": "10.0.0.2", 00:23:46.381 "adrfam": "ipv4", 00:23:46.381 "trsvcid": "4420", 00:23:46.381 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:23:46.381 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:23:46.381 "prchk_reftag": false, 00:23:46.381 "prchk_guard": false, 00:23:46.381 "hdgst": false, 00:23:46.381 "ddgst": false, 00:23:46.381 "psk": "key0", 00:23:46.381 "allow_unrecognized_csi": false, 00:23:46.381 "method": "bdev_nvme_attach_controller", 00:23:46.381 "req_id": 1 00:23:46.381 } 00:23:46.381 Got JSON-RPC error response 00:23:46.381 response: 00:23:46.381 { 00:23:46.381 "code": -126, 00:23:46.381 "message": "Required key not available" 00:23:46.381 } 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@37 -- # killprocess 1915388 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1915388 ']' 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1915388 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1915388 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1915388' 00:23:46.642 killing process with pid 1915388 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1915388 00:23:46.642 Received shutdown signal, test time was about 10.000000 seconds 00:23:46.642 00:23:46.642 Latency(us) 00:23:46.642 [2024-11-06T07:59:36.755Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:46.642 [2024-11-06T07:59:36.755Z] =================================================================================================================== 00:23:46.642 [2024-11-06T07:59:36.755Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1915388 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@38 -- # return 1 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:23:46.642 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@175 -- # killprocess 1912985 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1912985 ']' 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1912985 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1912985 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1912985' 00:23:46.643 killing process with pid 1912985 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1912985 00:23:46.643 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1912985 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@176 -- # nvmfappstart -m 0x2 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=1915699 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 1915699 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1915699 ']' 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:46.903 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:46.903 08:59:36 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:46.903 [2024-11-06 08:59:36.911944] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:46.903 [2024-11-06 08:59:36.912001] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:46.903 [2024-11-06 08:59:37.003067] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:47.194 [2024-11-06 08:59:37.032777] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:47.194 [2024-11-06 08:59:37.032811] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:47.194 [2024-11-06 08:59:37.032817] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:47.194 [2024-11-06 08:59:37.032821] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:47.194 [2024-11-06 08:59:37.032826] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:47.194 [2024-11-06 08:59:37.033316] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@178 -- # NOT setup_nvmf_tgt /tmp/tmp.mY106GdROb 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@650 -- # local es=0 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@652 -- # valid_exec_arg setup_nvmf_tgt /tmp/tmp.mY106GdROb 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@638 -- # local arg=setup_nvmf_tgt 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # type -t setup_nvmf_tgt 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # setup_nvmf_tgt /tmp/tmp.mY106GdROb 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.mY106GdROb 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:23:47.887 [2024-11-06 08:59:37.886043] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:47.887 08:59:37 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:23:48.148 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:23:48.148 [2024-11-06 08:59:38.198821] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:23:48.148 [2024-11-06 08:59:38.199020] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:48.148 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:23:48.408 malloc0 00:23:48.408 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:23:48.668 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.mY106GdROb 00:23:48.668 [2024-11-06 08:59:38.665664] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.mY106GdROb': 0100666 00:23:48.668 [2024-11-06 08:59:38.665683] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:23:48.668 request: 00:23:48.668 { 00:23:48.669 "name": "key0", 00:23:48.669 "path": "/tmp/tmp.mY106GdROb", 00:23:48.669 "method": "keyring_file_add_key", 00:23:48.669 "req_id": 1 00:23:48.669 } 00:23:48.669 Got JSON-RPC error response 00:23:48.669 response: 00:23:48.669 { 00:23:48.669 "code": -1, 00:23:48.669 "message": "Operation not permitted" 00:23:48.669 } 00:23:48.669 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:23:48.929 [2024-11-06 08:59:38.818062] tcp.c:3792:nvmf_tcp_subsystem_add_host: *ERROR*: Key 'key0' does not exist 00:23:48.929 [2024-11-06 08:59:38.818087] subsystem.c:1051:spdk_nvmf_subsystem_add_host_ext: *ERROR*: Unable to add host to TCP transport 00:23:48.929 request: 00:23:48.929 { 00:23:48.929 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:23:48.929 "host": "nqn.2016-06.io.spdk:host1", 00:23:48.929 "psk": "key0", 00:23:48.929 "method": "nvmf_subsystem_add_host", 00:23:48.929 "req_id": 1 00:23:48.929 } 00:23:48.929 Got JSON-RPC error response 00:23:48.929 response: 00:23:48.929 { 00:23:48.929 "code": -32603, 00:23:48.929 "message": "Internal error" 00:23:48.929 } 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@653 -- # es=1 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@181 -- # killprocess 1915699 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1915699 ']' 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1915699 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1915699 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1915699' 00:23:48.929 killing process with pid 1915699 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1915699 00:23:48.929 08:59:38 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1915699 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@182 -- # chmod 0600 /tmp/tmp.mY106GdROb 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@185 -- # nvmfappstart -m 0x2 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=1916079 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 1916079 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1916079 ']' 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:48.929 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:48.929 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:48.930 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:48.930 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:49.191 [2024-11-06 08:59:39.062496] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:49.191 [2024-11-06 08:59:39.062548] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:49.191 [2024-11-06 08:59:39.151961] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:49.191 [2024-11-06 08:59:39.180949] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:49.191 [2024-11-06 08:59:39.180980] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:49.191 [2024-11-06 08:59:39.180987] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:49.191 [2024-11-06 08:59:39.180992] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:49.191 [2024-11-06 08:59:39.180997] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:49.191 [2024-11-06 08:59:39.181456] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:49.761 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:49.761 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:49.761 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:23:49.761 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:23:49.761 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:50.021 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:50.021 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@186 -- # setup_nvmf_tgt /tmp/tmp.mY106GdROb 00:23:50.021 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.mY106GdROb 00:23:50.021 08:59:39 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:23:50.021 [2024-11-06 08:59:40.042137] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:50.021 08:59:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:23:50.281 08:59:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:23:50.281 [2024-11-06 08:59:40.366926] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:23:50.281 [2024-11-06 08:59:40.367119] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:50.281 08:59:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:23:50.541 malloc0 00:23:50.541 08:59:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:23:50.801 08:59:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.mY106GdROb 00:23:50.801 08:59:40 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:23:51.062 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@189 -- # bdevperf_pid=1916465 00:23:51.062 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@191 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:23:51.062 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@188 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:23:51.062 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@192 -- # waitforlisten 1916465 /var/tmp/bdevperf.sock 00:23:51.062 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1916465 ']' 00:23:51.062 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:51.062 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:51.062 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:51.062 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:51.062 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:51.062 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:51.062 [2024-11-06 08:59:41.055927] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:51.062 [2024-11-06 08:59:41.055984] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1916465 ] 00:23:51.062 [2024-11-06 08:59:41.114841] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:51.062 [2024-11-06 08:59:41.144140] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:51.323 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:51.323 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:51.323 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@193 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.mY106GdROb 00:23:51.324 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@194 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:23:51.585 [2024-11-06 08:59:41.562387] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:23:51.585 TLSTESTn1 00:23:51.585 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@198 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py save_config 00:23:51.847 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@198 -- # tgtconf='{ 00:23:51.847 "subsystems": [ 00:23:51.847 { 00:23:51.847 "subsystem": "keyring", 00:23:51.847 "config": [ 00:23:51.847 { 00:23:51.847 "method": "keyring_file_add_key", 00:23:51.847 "params": { 00:23:51.847 "name": "key0", 00:23:51.847 "path": "/tmp/tmp.mY106GdROb" 00:23:51.847 } 00:23:51.847 } 00:23:51.847 ] 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "subsystem": "iobuf", 00:23:51.847 "config": [ 00:23:51.847 { 00:23:51.847 "method": "iobuf_set_options", 00:23:51.847 "params": { 00:23:51.847 "small_pool_count": 8192, 00:23:51.847 "large_pool_count": 1024, 00:23:51.847 "small_bufsize": 8192, 00:23:51.847 "large_bufsize": 135168, 00:23:51.847 "enable_numa": false 00:23:51.847 } 00:23:51.847 } 00:23:51.847 ] 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "subsystem": "sock", 00:23:51.847 "config": [ 00:23:51.847 { 00:23:51.847 "method": "sock_set_default_impl", 00:23:51.847 "params": { 00:23:51.847 "impl_name": "posix" 00:23:51.847 } 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "method": "sock_impl_set_options", 00:23:51.847 "params": { 00:23:51.847 "impl_name": "ssl", 00:23:51.847 "recv_buf_size": 4096, 00:23:51.847 "send_buf_size": 4096, 00:23:51.847 "enable_recv_pipe": true, 00:23:51.847 "enable_quickack": false, 00:23:51.847 "enable_placement_id": 0, 00:23:51.847 "enable_zerocopy_send_server": true, 00:23:51.847 "enable_zerocopy_send_client": false, 00:23:51.847 "zerocopy_threshold": 0, 00:23:51.847 "tls_version": 0, 00:23:51.847 "enable_ktls": false 00:23:51.847 } 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "method": "sock_impl_set_options", 00:23:51.847 "params": { 00:23:51.847 "impl_name": "posix", 00:23:51.847 "recv_buf_size": 2097152, 00:23:51.847 "send_buf_size": 2097152, 00:23:51.847 "enable_recv_pipe": true, 00:23:51.847 "enable_quickack": false, 00:23:51.847 "enable_placement_id": 0, 00:23:51.847 "enable_zerocopy_send_server": true, 00:23:51.847 "enable_zerocopy_send_client": false, 00:23:51.847 "zerocopy_threshold": 0, 00:23:51.847 "tls_version": 0, 00:23:51.847 "enable_ktls": false 00:23:51.847 } 00:23:51.847 } 00:23:51.847 ] 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "subsystem": "vmd", 00:23:51.847 "config": [] 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "subsystem": "accel", 00:23:51.847 "config": [ 00:23:51.847 { 00:23:51.847 "method": "accel_set_options", 00:23:51.847 "params": { 00:23:51.847 "small_cache_size": 128, 00:23:51.847 "large_cache_size": 16, 00:23:51.847 "task_count": 2048, 00:23:51.847 "sequence_count": 2048, 00:23:51.847 "buf_count": 2048 00:23:51.847 } 00:23:51.847 } 00:23:51.847 ] 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "subsystem": "bdev", 00:23:51.847 "config": [ 00:23:51.847 { 00:23:51.847 "method": "bdev_set_options", 00:23:51.847 "params": { 00:23:51.847 "bdev_io_pool_size": 65535, 00:23:51.847 "bdev_io_cache_size": 256, 00:23:51.847 "bdev_auto_examine": true, 00:23:51.847 "iobuf_small_cache_size": 128, 00:23:51.847 "iobuf_large_cache_size": 16 00:23:51.847 } 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "method": "bdev_raid_set_options", 00:23:51.847 "params": { 00:23:51.847 "process_window_size_kb": 1024, 00:23:51.847 "process_max_bandwidth_mb_sec": 0 00:23:51.847 } 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "method": "bdev_iscsi_set_options", 00:23:51.847 "params": { 00:23:51.847 "timeout_sec": 30 00:23:51.847 } 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "method": "bdev_nvme_set_options", 00:23:51.847 "params": { 00:23:51.847 "action_on_timeout": "none", 00:23:51.847 "timeout_us": 0, 00:23:51.847 "timeout_admin_us": 0, 00:23:51.847 "keep_alive_timeout_ms": 10000, 00:23:51.847 "arbitration_burst": 0, 00:23:51.847 "low_priority_weight": 0, 00:23:51.847 "medium_priority_weight": 0, 00:23:51.847 "high_priority_weight": 0, 00:23:51.847 "nvme_adminq_poll_period_us": 10000, 00:23:51.847 "nvme_ioq_poll_period_us": 0, 00:23:51.847 "io_queue_requests": 0, 00:23:51.847 "delay_cmd_submit": true, 00:23:51.847 "transport_retry_count": 4, 00:23:51.847 "bdev_retry_count": 3, 00:23:51.847 "transport_ack_timeout": 0, 00:23:51.847 "ctrlr_loss_timeout_sec": 0, 00:23:51.847 "reconnect_delay_sec": 0, 00:23:51.847 "fast_io_fail_timeout_sec": 0, 00:23:51.847 "disable_auto_failback": false, 00:23:51.847 "generate_uuids": false, 00:23:51.847 "transport_tos": 0, 00:23:51.847 "nvme_error_stat": false, 00:23:51.847 "rdma_srq_size": 0, 00:23:51.847 "io_path_stat": false, 00:23:51.847 "allow_accel_sequence": false, 00:23:51.847 "rdma_max_cq_size": 0, 00:23:51.847 "rdma_cm_event_timeout_ms": 0, 00:23:51.847 "dhchap_digests": [ 00:23:51.847 "sha256", 00:23:51.847 "sha384", 00:23:51.847 "sha512" 00:23:51.847 ], 00:23:51.847 "dhchap_dhgroups": [ 00:23:51.847 "null", 00:23:51.847 "ffdhe2048", 00:23:51.847 "ffdhe3072", 00:23:51.847 "ffdhe4096", 00:23:51.847 "ffdhe6144", 00:23:51.847 "ffdhe8192" 00:23:51.847 ] 00:23:51.847 } 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "method": "bdev_nvme_set_hotplug", 00:23:51.847 "params": { 00:23:51.847 "period_us": 100000, 00:23:51.847 "enable": false 00:23:51.847 } 00:23:51.847 }, 00:23:51.847 { 00:23:51.847 "method": "bdev_malloc_create", 00:23:51.847 "params": { 00:23:51.847 "name": "malloc0", 00:23:51.847 "num_blocks": 8192, 00:23:51.847 "block_size": 4096, 00:23:51.847 "physical_block_size": 4096, 00:23:51.847 "uuid": "26281fa4-32e8-49f7-a4e7-0e86cfe7f5f1", 00:23:51.847 "optimal_io_boundary": 0, 00:23:51.847 "md_size": 0, 00:23:51.847 "dif_type": 0, 00:23:51.847 "dif_is_head_of_md": false, 00:23:51.847 "dif_pi_format": 0 00:23:51.847 } 00:23:51.847 }, 00:23:51.847 { 00:23:51.848 "method": "bdev_wait_for_examine" 00:23:51.848 } 00:23:51.848 ] 00:23:51.848 }, 00:23:51.848 { 00:23:51.848 "subsystem": "nbd", 00:23:51.848 "config": [] 00:23:51.848 }, 00:23:51.848 { 00:23:51.848 "subsystem": "scheduler", 00:23:51.848 "config": [ 00:23:51.848 { 00:23:51.848 "method": "framework_set_scheduler", 00:23:51.848 "params": { 00:23:51.848 "name": "static" 00:23:51.848 } 00:23:51.848 } 00:23:51.848 ] 00:23:51.848 }, 00:23:51.848 { 00:23:51.848 "subsystem": "nvmf", 00:23:51.848 "config": [ 00:23:51.848 { 00:23:51.848 "method": "nvmf_set_config", 00:23:51.848 "params": { 00:23:51.848 "discovery_filter": "match_any", 00:23:51.848 "admin_cmd_passthru": { 00:23:51.848 "identify_ctrlr": false 00:23:51.848 }, 00:23:51.848 "dhchap_digests": [ 00:23:51.848 "sha256", 00:23:51.848 "sha384", 00:23:51.848 "sha512" 00:23:51.848 ], 00:23:51.848 "dhchap_dhgroups": [ 00:23:51.848 "null", 00:23:51.848 "ffdhe2048", 00:23:51.848 "ffdhe3072", 00:23:51.848 "ffdhe4096", 00:23:51.848 "ffdhe6144", 00:23:51.848 "ffdhe8192" 00:23:51.848 ] 00:23:51.848 } 00:23:51.848 }, 00:23:51.848 { 00:23:51.848 "method": "nvmf_set_max_subsystems", 00:23:51.848 "params": { 00:23:51.848 "max_subsystems": 1024 00:23:51.848 } 00:23:51.848 }, 00:23:51.848 { 00:23:51.848 "method": "nvmf_set_crdt", 00:23:51.848 "params": { 00:23:51.848 "crdt1": 0, 00:23:51.848 "crdt2": 0, 00:23:51.848 "crdt3": 0 00:23:51.848 } 00:23:51.848 }, 00:23:51.848 { 00:23:51.848 "method": "nvmf_create_transport", 00:23:51.848 "params": { 00:23:51.848 "trtype": "TCP", 00:23:51.848 "max_queue_depth": 128, 00:23:51.848 "max_io_qpairs_per_ctrlr": 127, 00:23:51.848 "in_capsule_data_size": 4096, 00:23:51.848 "max_io_size": 131072, 00:23:51.848 "io_unit_size": 131072, 00:23:51.848 "max_aq_depth": 128, 00:23:51.848 "num_shared_buffers": 511, 00:23:51.848 "buf_cache_size": 4294967295, 00:23:51.848 "dif_insert_or_strip": false, 00:23:51.848 "zcopy": false, 00:23:51.848 "c2h_success": false, 00:23:51.848 "sock_priority": 0, 00:23:51.848 "abort_timeout_sec": 1, 00:23:51.848 "ack_timeout": 0, 00:23:51.848 "data_wr_pool_size": 0 00:23:51.848 } 00:23:51.848 }, 00:23:51.848 { 00:23:51.848 "method": "nvmf_create_subsystem", 00:23:51.848 "params": { 00:23:51.848 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:23:51.848 "allow_any_host": false, 00:23:51.848 "serial_number": "SPDK00000000000001", 00:23:51.848 "model_number": "SPDK bdev Controller", 00:23:51.848 "max_namespaces": 10, 00:23:51.848 "min_cntlid": 1, 00:23:51.848 "max_cntlid": 65519, 00:23:51.848 "ana_reporting": false 00:23:51.848 } 00:23:51.848 }, 00:23:51.848 { 00:23:51.848 "method": "nvmf_subsystem_add_host", 00:23:51.848 "params": { 00:23:51.848 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:23:51.848 "host": "nqn.2016-06.io.spdk:host1", 00:23:51.848 "psk": "key0" 00:23:51.848 } 00:23:51.848 }, 00:23:51.848 { 00:23:51.848 "method": "nvmf_subsystem_add_ns", 00:23:51.848 "params": { 00:23:51.848 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:23:51.848 "namespace": { 00:23:51.848 "nsid": 1, 00:23:51.848 "bdev_name": "malloc0", 00:23:51.848 "nguid": "26281FA432E849F7A4E70E86CFE7F5F1", 00:23:51.848 "uuid": "26281fa4-32e8-49f7-a4e7-0e86cfe7f5f1", 00:23:51.848 "no_auto_visible": false 00:23:51.848 } 00:23:51.848 } 00:23:51.848 }, 00:23:51.848 { 00:23:51.848 "method": "nvmf_subsystem_add_listener", 00:23:51.848 "params": { 00:23:51.848 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:23:51.848 "listen_address": { 00:23:51.848 "trtype": "TCP", 00:23:51.848 "adrfam": "IPv4", 00:23:51.848 "traddr": "10.0.0.2", 00:23:51.848 "trsvcid": "4420" 00:23:51.848 }, 00:23:51.848 "secure_channel": true 00:23:51.848 } 00:23:51.848 } 00:23:51.848 ] 00:23:51.848 } 00:23:51.848 ] 00:23:51.848 }' 00:23:51.848 08:59:41 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@199 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:23:52.109 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@199 -- # bdevperfconf='{ 00:23:52.109 "subsystems": [ 00:23:52.109 { 00:23:52.109 "subsystem": "keyring", 00:23:52.109 "config": [ 00:23:52.109 { 00:23:52.109 "method": "keyring_file_add_key", 00:23:52.109 "params": { 00:23:52.110 "name": "key0", 00:23:52.110 "path": "/tmp/tmp.mY106GdROb" 00:23:52.110 } 00:23:52.110 } 00:23:52.110 ] 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "subsystem": "iobuf", 00:23:52.110 "config": [ 00:23:52.110 { 00:23:52.110 "method": "iobuf_set_options", 00:23:52.110 "params": { 00:23:52.110 "small_pool_count": 8192, 00:23:52.110 "large_pool_count": 1024, 00:23:52.110 "small_bufsize": 8192, 00:23:52.110 "large_bufsize": 135168, 00:23:52.110 "enable_numa": false 00:23:52.110 } 00:23:52.110 } 00:23:52.110 ] 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "subsystem": "sock", 00:23:52.110 "config": [ 00:23:52.110 { 00:23:52.110 "method": "sock_set_default_impl", 00:23:52.110 "params": { 00:23:52.110 "impl_name": "posix" 00:23:52.110 } 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "method": "sock_impl_set_options", 00:23:52.110 "params": { 00:23:52.110 "impl_name": "ssl", 00:23:52.110 "recv_buf_size": 4096, 00:23:52.110 "send_buf_size": 4096, 00:23:52.110 "enable_recv_pipe": true, 00:23:52.110 "enable_quickack": false, 00:23:52.110 "enable_placement_id": 0, 00:23:52.110 "enable_zerocopy_send_server": true, 00:23:52.110 "enable_zerocopy_send_client": false, 00:23:52.110 "zerocopy_threshold": 0, 00:23:52.110 "tls_version": 0, 00:23:52.110 "enable_ktls": false 00:23:52.110 } 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "method": "sock_impl_set_options", 00:23:52.110 "params": { 00:23:52.110 "impl_name": "posix", 00:23:52.110 "recv_buf_size": 2097152, 00:23:52.110 "send_buf_size": 2097152, 00:23:52.110 "enable_recv_pipe": true, 00:23:52.110 "enable_quickack": false, 00:23:52.110 "enable_placement_id": 0, 00:23:52.110 "enable_zerocopy_send_server": true, 00:23:52.110 "enable_zerocopy_send_client": false, 00:23:52.110 "zerocopy_threshold": 0, 00:23:52.110 "tls_version": 0, 00:23:52.110 "enable_ktls": false 00:23:52.110 } 00:23:52.110 } 00:23:52.110 ] 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "subsystem": "vmd", 00:23:52.110 "config": [] 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "subsystem": "accel", 00:23:52.110 "config": [ 00:23:52.110 { 00:23:52.110 "method": "accel_set_options", 00:23:52.110 "params": { 00:23:52.110 "small_cache_size": 128, 00:23:52.110 "large_cache_size": 16, 00:23:52.110 "task_count": 2048, 00:23:52.110 "sequence_count": 2048, 00:23:52.110 "buf_count": 2048 00:23:52.110 } 00:23:52.110 } 00:23:52.110 ] 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "subsystem": "bdev", 00:23:52.110 "config": [ 00:23:52.110 { 00:23:52.110 "method": "bdev_set_options", 00:23:52.110 "params": { 00:23:52.110 "bdev_io_pool_size": 65535, 00:23:52.110 "bdev_io_cache_size": 256, 00:23:52.110 "bdev_auto_examine": true, 00:23:52.110 "iobuf_small_cache_size": 128, 00:23:52.110 "iobuf_large_cache_size": 16 00:23:52.110 } 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "method": "bdev_raid_set_options", 00:23:52.110 "params": { 00:23:52.110 "process_window_size_kb": 1024, 00:23:52.110 "process_max_bandwidth_mb_sec": 0 00:23:52.110 } 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "method": "bdev_iscsi_set_options", 00:23:52.110 "params": { 00:23:52.110 "timeout_sec": 30 00:23:52.110 } 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "method": "bdev_nvme_set_options", 00:23:52.110 "params": { 00:23:52.110 "action_on_timeout": "none", 00:23:52.110 "timeout_us": 0, 00:23:52.110 "timeout_admin_us": 0, 00:23:52.110 "keep_alive_timeout_ms": 10000, 00:23:52.110 "arbitration_burst": 0, 00:23:52.110 "low_priority_weight": 0, 00:23:52.110 "medium_priority_weight": 0, 00:23:52.110 "high_priority_weight": 0, 00:23:52.110 "nvme_adminq_poll_period_us": 10000, 00:23:52.110 "nvme_ioq_poll_period_us": 0, 00:23:52.110 "io_queue_requests": 512, 00:23:52.110 "delay_cmd_submit": true, 00:23:52.110 "transport_retry_count": 4, 00:23:52.110 "bdev_retry_count": 3, 00:23:52.110 "transport_ack_timeout": 0, 00:23:52.110 "ctrlr_loss_timeout_sec": 0, 00:23:52.110 "reconnect_delay_sec": 0, 00:23:52.110 "fast_io_fail_timeout_sec": 0, 00:23:52.110 "disable_auto_failback": false, 00:23:52.110 "generate_uuids": false, 00:23:52.110 "transport_tos": 0, 00:23:52.110 "nvme_error_stat": false, 00:23:52.110 "rdma_srq_size": 0, 00:23:52.110 "io_path_stat": false, 00:23:52.110 "allow_accel_sequence": false, 00:23:52.110 "rdma_max_cq_size": 0, 00:23:52.110 "rdma_cm_event_timeout_ms": 0, 00:23:52.110 "dhchap_digests": [ 00:23:52.110 "sha256", 00:23:52.110 "sha384", 00:23:52.110 "sha512" 00:23:52.110 ], 00:23:52.110 "dhchap_dhgroups": [ 00:23:52.110 "null", 00:23:52.110 "ffdhe2048", 00:23:52.110 "ffdhe3072", 00:23:52.110 "ffdhe4096", 00:23:52.110 "ffdhe6144", 00:23:52.110 "ffdhe8192" 00:23:52.110 ] 00:23:52.110 } 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "method": "bdev_nvme_attach_controller", 00:23:52.110 "params": { 00:23:52.110 "name": "TLSTEST", 00:23:52.110 "trtype": "TCP", 00:23:52.110 "adrfam": "IPv4", 00:23:52.110 "traddr": "10.0.0.2", 00:23:52.110 "trsvcid": "4420", 00:23:52.110 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:23:52.110 "prchk_reftag": false, 00:23:52.110 "prchk_guard": false, 00:23:52.110 "ctrlr_loss_timeout_sec": 0, 00:23:52.110 "reconnect_delay_sec": 0, 00:23:52.110 "fast_io_fail_timeout_sec": 0, 00:23:52.110 "psk": "key0", 00:23:52.110 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:23:52.110 "hdgst": false, 00:23:52.110 "ddgst": false, 00:23:52.110 "multipath": "multipath" 00:23:52.110 } 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "method": "bdev_nvme_set_hotplug", 00:23:52.110 "params": { 00:23:52.110 "period_us": 100000, 00:23:52.110 "enable": false 00:23:52.110 } 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "method": "bdev_wait_for_examine" 00:23:52.110 } 00:23:52.110 ] 00:23:52.110 }, 00:23:52.110 { 00:23:52.110 "subsystem": "nbd", 00:23:52.110 "config": [] 00:23:52.110 } 00:23:52.110 ] 00:23:52.110 }' 00:23:52.110 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@201 -- # killprocess 1916465 00:23:52.110 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1916465 ']' 00:23:52.110 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1916465 00:23:52.110 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:52.110 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:52.110 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1916465 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1916465' 00:23:52.372 killing process with pid 1916465 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1916465 00:23:52.372 Received shutdown signal, test time was about 10.000000 seconds 00:23:52.372 00:23:52.372 Latency(us) 00:23:52.372 [2024-11-06T07:59:42.485Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:52.372 [2024-11-06T07:59:42.485Z] =================================================================================================================== 00:23:52.372 [2024-11-06T07:59:42.485Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1916465 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@202 -- # killprocess 1916079 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1916079 ']' 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1916079 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1916079 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1916079' 00:23:52.372 killing process with pid 1916079 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1916079 00:23:52.372 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1916079 00:23:52.633 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@205 -- # nvmfappstart -m 0x2 -c /dev/fd/62 00:23:52.633 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:23:52.633 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:23:52.633 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:52.633 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@205 -- # echo '{ 00:23:52.633 "subsystems": [ 00:23:52.633 { 00:23:52.633 "subsystem": "keyring", 00:23:52.633 "config": [ 00:23:52.633 { 00:23:52.633 "method": "keyring_file_add_key", 00:23:52.633 "params": { 00:23:52.633 "name": "key0", 00:23:52.633 "path": "/tmp/tmp.mY106GdROb" 00:23:52.633 } 00:23:52.633 } 00:23:52.633 ] 00:23:52.633 }, 00:23:52.633 { 00:23:52.633 "subsystem": "iobuf", 00:23:52.633 "config": [ 00:23:52.633 { 00:23:52.633 "method": "iobuf_set_options", 00:23:52.633 "params": { 00:23:52.633 "small_pool_count": 8192, 00:23:52.633 "large_pool_count": 1024, 00:23:52.633 "small_bufsize": 8192, 00:23:52.633 "large_bufsize": 135168, 00:23:52.633 "enable_numa": false 00:23:52.633 } 00:23:52.633 } 00:23:52.633 ] 00:23:52.633 }, 00:23:52.633 { 00:23:52.633 "subsystem": "sock", 00:23:52.633 "config": [ 00:23:52.633 { 00:23:52.633 "method": "sock_set_default_impl", 00:23:52.633 "params": { 00:23:52.633 "impl_name": "posix" 00:23:52.634 } 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "method": "sock_impl_set_options", 00:23:52.634 "params": { 00:23:52.634 "impl_name": "ssl", 00:23:52.634 "recv_buf_size": 4096, 00:23:52.634 "send_buf_size": 4096, 00:23:52.634 "enable_recv_pipe": true, 00:23:52.634 "enable_quickack": false, 00:23:52.634 "enable_placement_id": 0, 00:23:52.634 "enable_zerocopy_send_server": true, 00:23:52.634 "enable_zerocopy_send_client": false, 00:23:52.634 "zerocopy_threshold": 0, 00:23:52.634 "tls_version": 0, 00:23:52.634 "enable_ktls": false 00:23:52.634 } 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "method": "sock_impl_set_options", 00:23:52.634 "params": { 00:23:52.634 "impl_name": "posix", 00:23:52.634 "recv_buf_size": 2097152, 00:23:52.634 "send_buf_size": 2097152, 00:23:52.634 "enable_recv_pipe": true, 00:23:52.634 "enable_quickack": false, 00:23:52.634 "enable_placement_id": 0, 00:23:52.634 "enable_zerocopy_send_server": true, 00:23:52.634 "enable_zerocopy_send_client": false, 00:23:52.634 "zerocopy_threshold": 0, 00:23:52.634 "tls_version": 0, 00:23:52.634 "enable_ktls": false 00:23:52.634 } 00:23:52.634 } 00:23:52.634 ] 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "subsystem": "vmd", 00:23:52.634 "config": [] 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "subsystem": "accel", 00:23:52.634 "config": [ 00:23:52.634 { 00:23:52.634 "method": "accel_set_options", 00:23:52.634 "params": { 00:23:52.634 "small_cache_size": 128, 00:23:52.634 "large_cache_size": 16, 00:23:52.634 "task_count": 2048, 00:23:52.634 "sequence_count": 2048, 00:23:52.634 "buf_count": 2048 00:23:52.634 } 00:23:52.634 } 00:23:52.634 ] 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "subsystem": "bdev", 00:23:52.634 "config": [ 00:23:52.634 { 00:23:52.634 "method": "bdev_set_options", 00:23:52.634 "params": { 00:23:52.634 "bdev_io_pool_size": 65535, 00:23:52.634 "bdev_io_cache_size": 256, 00:23:52.634 "bdev_auto_examine": true, 00:23:52.634 "iobuf_small_cache_size": 128, 00:23:52.634 "iobuf_large_cache_size": 16 00:23:52.634 } 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "method": "bdev_raid_set_options", 00:23:52.634 "params": { 00:23:52.634 "process_window_size_kb": 1024, 00:23:52.634 "process_max_bandwidth_mb_sec": 0 00:23:52.634 } 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "method": "bdev_iscsi_set_options", 00:23:52.634 "params": { 00:23:52.634 "timeout_sec": 30 00:23:52.634 } 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "method": "bdev_nvme_set_options", 00:23:52.634 "params": { 00:23:52.634 "action_on_timeout": "none", 00:23:52.634 "timeout_us": 0, 00:23:52.634 "timeout_admin_us": 0, 00:23:52.634 "keep_alive_timeout_ms": 10000, 00:23:52.634 "arbitration_burst": 0, 00:23:52.634 "low_priority_weight": 0, 00:23:52.634 "medium_priority_weight": 0, 00:23:52.634 "high_priority_weight": 0, 00:23:52.634 "nvme_adminq_poll_period_us": 10000, 00:23:52.634 "nvme_ioq_poll_period_us": 0, 00:23:52.634 "io_queue_requests": 0, 00:23:52.634 "delay_cmd_submit": true, 00:23:52.634 "transport_retry_count": 4, 00:23:52.634 "bdev_retry_count": 3, 00:23:52.634 "transport_ack_timeout": 0, 00:23:52.634 "ctrlr_loss_timeout_sec": 0, 00:23:52.634 "reconnect_delay_sec": 0, 00:23:52.634 "fast_io_fail_timeout_sec": 0, 00:23:52.634 "disable_auto_failback": false, 00:23:52.634 "generate_uuids": false, 00:23:52.634 "transport_tos": 0, 00:23:52.634 "nvme_error_stat": false, 00:23:52.634 "rdma_srq_size": 0, 00:23:52.634 "io_path_stat": false, 00:23:52.634 "allow_accel_sequence": false, 00:23:52.634 "rdma_max_cq_size": 0, 00:23:52.634 "rdma_cm_event_timeout_ms": 0, 00:23:52.634 "dhchap_digests": [ 00:23:52.634 "sha256", 00:23:52.634 "sha384", 00:23:52.634 "sha512" 00:23:52.634 ], 00:23:52.634 "dhchap_dhgroups": [ 00:23:52.634 "null", 00:23:52.634 "ffdhe2048", 00:23:52.634 "ffdhe3072", 00:23:52.634 "ffdhe4096", 00:23:52.634 "ffdhe6144", 00:23:52.634 "ffdhe8192" 00:23:52.634 ] 00:23:52.634 } 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "method": "bdev_nvme_set_hotplug", 00:23:52.634 "params": { 00:23:52.634 "period_us": 100000, 00:23:52.634 "enable": false 00:23:52.634 } 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "method": "bdev_malloc_create", 00:23:52.634 "params": { 00:23:52.634 "name": "malloc0", 00:23:52.634 "num_blocks": 8192, 00:23:52.634 "block_size": 4096, 00:23:52.634 "physical_block_size": 4096, 00:23:52.634 "uuid": "26281fa4-32e8-49f7-a4e7-0e86cfe7f5f1", 00:23:52.634 "optimal_io_boundary": 0, 00:23:52.634 "md_size": 0, 00:23:52.634 "dif_type": 0, 00:23:52.634 "dif_is_head_of_md": false, 00:23:52.634 "dif_pi_format": 0 00:23:52.634 } 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "method": "bdev_wait_for_examine" 00:23:52.634 } 00:23:52.634 ] 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "subsystem": "nbd", 00:23:52.634 "config": [] 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "subsystem": "scheduler", 00:23:52.634 "config": [ 00:23:52.634 { 00:23:52.634 "method": "framework_set_scheduler", 00:23:52.634 "params": { 00:23:52.634 "name": "static" 00:23:52.634 } 00:23:52.634 } 00:23:52.634 ] 00:23:52.634 }, 00:23:52.634 { 00:23:52.634 "subsystem": "nvmf", 00:23:52.634 "config": [ 00:23:52.634 { 00:23:52.634 "method": "nvmf_set_config", 00:23:52.634 "params": { 00:23:52.634 "discovery_filter": "match_any", 00:23:52.634 "admin_cmd_passthru": { 00:23:52.634 "identify_ctrlr": false 00:23:52.634 }, 00:23:52.634 "dhchap_digests": [ 00:23:52.634 "sha256", 00:23:52.634 "sha384", 00:23:52.634 "sha512" 00:23:52.635 ], 00:23:52.635 "dhchap_dhgroups": [ 00:23:52.635 "null", 00:23:52.635 "ffdhe2048", 00:23:52.635 "ffdhe3072", 00:23:52.635 "ffdhe4096", 00:23:52.635 "ffdhe6144", 00:23:52.635 "ffdhe8192" 00:23:52.635 ] 00:23:52.635 } 00:23:52.635 }, 00:23:52.635 { 00:23:52.635 "method": "nvmf_set_max_subsystems", 00:23:52.635 "params": { 00:23:52.635 "max_subsystems": 1024 00:23:52.635 } 00:23:52.635 }, 00:23:52.635 { 00:23:52.635 "method": "nvmf_set_crdt", 00:23:52.635 "params": { 00:23:52.635 "crdt1": 0, 00:23:52.635 "crdt2": 0, 00:23:52.635 "crdt3": 0 00:23:52.635 } 00:23:52.635 }, 00:23:52.635 { 00:23:52.635 "method": "nvmf_create_transport", 00:23:52.635 "params": { 00:23:52.635 "trtype": "TCP", 00:23:52.635 "max_queue_depth": 128, 00:23:52.635 "max_io_qpairs_per_ctrlr": 127, 00:23:52.635 "in_capsule_data_size": 4096, 00:23:52.635 "max_io_size": 131072, 00:23:52.635 "io_unit_size": 131072, 00:23:52.635 "max_aq_depth": 128, 00:23:52.635 "num_shared_buffers": 511, 00:23:52.635 "buf_cache_size": 4294967295, 00:23:52.635 "dif_insert_or_strip": false, 00:23:52.635 "zcopy": false, 00:23:52.635 "c2h_success": false, 00:23:52.635 "sock_priority": 0, 00:23:52.635 "abort_timeout_sec": 1, 00:23:52.635 "ack_timeout": 0, 00:23:52.635 "data_wr_pool_size": 0 00:23:52.635 } 00:23:52.635 }, 00:23:52.635 { 00:23:52.635 "method": "nvmf_create_subsystem", 00:23:52.635 "params": { 00:23:52.635 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:23:52.635 "allow_any_host": false, 00:23:52.635 "serial_number": "SPDK00000000000001", 00:23:52.635 "model_number": "SPDK bdev Controller", 00:23:52.635 "max_namespaces": 10, 00:23:52.635 "min_cntlid": 1, 00:23:52.635 "max_cntlid": 65519, 00:23:52.635 "ana_reporting": false 00:23:52.635 } 00:23:52.635 }, 00:23:52.635 { 00:23:52.635 "method": "nvmf_subsystem_add_host", 00:23:52.635 "params": { 00:23:52.635 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:23:52.635 "host": "nqn.2016-06.io.spdk:host1", 00:23:52.635 "psk": "key0" 00:23:52.635 } 00:23:52.635 }, 00:23:52.635 { 00:23:52.635 "method": "nvmf_subsystem_add_ns", 00:23:52.635 "params": { 00:23:52.635 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:23:52.635 "namespace": { 00:23:52.635 "nsid": 1, 00:23:52.635 "bdev_name": "malloc0", 00:23:52.635 "nguid": "26281FA432E849F7A4E70E86CFE7F5F1", 00:23:52.635 "uuid": "26281fa4-32e8-49f7-a4e7-0e86cfe7f5f1", 00:23:52.635 "no_auto_visible": false 00:23:52.635 } 00:23:52.635 } 00:23:52.635 }, 00:23:52.635 { 00:23:52.635 "method": "nvmf_subsystem_add_listener", 00:23:52.635 "params": { 00:23:52.635 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:23:52.635 "listen_address": { 00:23:52.635 "trtype": "TCP", 00:23:52.635 "adrfam": "IPv4", 00:23:52.635 "traddr": "10.0.0.2", 00:23:52.635 "trsvcid": "4420" 00:23:52.635 }, 00:23:52.635 "secure_channel": true 00:23:52.635 } 00:23:52.635 } 00:23:52.635 ] 00:23:52.635 } 00:23:52.635 ] 00:23:52.635 }' 00:23:52.635 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=1916800 00:23:52.635 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 1916800 00:23:52.635 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 -c /dev/fd/62 00:23:52.635 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1916800 ']' 00:23:52.635 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:52.635 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:52.635 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:52.635 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:52.635 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:52.635 08:59:42 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:52.635 [2024-11-06 08:59:42.569097] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:52.635 [2024-11-06 08:59:42.569154] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:52.635 [2024-11-06 08:59:42.660710] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:52.635 [2024-11-06 08:59:42.690433] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:52.635 [2024-11-06 08:59:42.690462] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:52.635 [2024-11-06 08:59:42.690468] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:52.635 [2024-11-06 08:59:42.690473] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:52.635 [2024-11-06 08:59:42.690477] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:52.635 [2024-11-06 08:59:42.690968] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:23:52.896 [2024-11-06 08:59:42.884691] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:52.896 [2024-11-06 08:59:42.916721] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:23:52.896 [2024-11-06 08:59:42.916915] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@209 -- # bdevperf_pid=1917145 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@210 -- # waitforlisten 1917145 /var/tmp/bdevperf.sock 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1917145 ']' 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:53.469 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@206 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 -c /dev/fd/63 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:23:53.469 08:59:43 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@206 -- # echo '{ 00:23:53.469 "subsystems": [ 00:23:53.469 { 00:23:53.469 "subsystem": "keyring", 00:23:53.469 "config": [ 00:23:53.469 { 00:23:53.469 "method": "keyring_file_add_key", 00:23:53.469 "params": { 00:23:53.469 "name": "key0", 00:23:53.469 "path": "/tmp/tmp.mY106GdROb" 00:23:53.469 } 00:23:53.469 } 00:23:53.469 ] 00:23:53.469 }, 00:23:53.469 { 00:23:53.469 "subsystem": "iobuf", 00:23:53.469 "config": [ 00:23:53.469 { 00:23:53.469 "method": "iobuf_set_options", 00:23:53.469 "params": { 00:23:53.469 "small_pool_count": 8192, 00:23:53.469 "large_pool_count": 1024, 00:23:53.469 "small_bufsize": 8192, 00:23:53.469 "large_bufsize": 135168, 00:23:53.469 "enable_numa": false 00:23:53.469 } 00:23:53.469 } 00:23:53.469 ] 00:23:53.469 }, 00:23:53.469 { 00:23:53.469 "subsystem": "sock", 00:23:53.469 "config": [ 00:23:53.469 { 00:23:53.469 "method": "sock_set_default_impl", 00:23:53.469 "params": { 00:23:53.469 "impl_name": "posix" 00:23:53.469 } 00:23:53.469 }, 00:23:53.469 { 00:23:53.469 "method": "sock_impl_set_options", 00:23:53.469 "params": { 00:23:53.469 "impl_name": "ssl", 00:23:53.469 "recv_buf_size": 4096, 00:23:53.469 "send_buf_size": 4096, 00:23:53.469 "enable_recv_pipe": true, 00:23:53.469 "enable_quickack": false, 00:23:53.469 "enable_placement_id": 0, 00:23:53.469 "enable_zerocopy_send_server": true, 00:23:53.469 "enable_zerocopy_send_client": false, 00:23:53.469 "zerocopy_threshold": 0, 00:23:53.469 "tls_version": 0, 00:23:53.469 "enable_ktls": false 00:23:53.469 } 00:23:53.469 }, 00:23:53.469 { 00:23:53.469 "method": "sock_impl_set_options", 00:23:53.469 "params": { 00:23:53.469 "impl_name": "posix", 00:23:53.469 "recv_buf_size": 2097152, 00:23:53.469 "send_buf_size": 2097152, 00:23:53.469 "enable_recv_pipe": true, 00:23:53.469 "enable_quickack": false, 00:23:53.469 "enable_placement_id": 0, 00:23:53.469 "enable_zerocopy_send_server": true, 00:23:53.469 "enable_zerocopy_send_client": false, 00:23:53.469 "zerocopy_threshold": 0, 00:23:53.469 "tls_version": 0, 00:23:53.469 "enable_ktls": false 00:23:53.469 } 00:23:53.469 } 00:23:53.469 ] 00:23:53.469 }, 00:23:53.469 { 00:23:53.469 "subsystem": "vmd", 00:23:53.469 "config": [] 00:23:53.469 }, 00:23:53.469 { 00:23:53.469 "subsystem": "accel", 00:23:53.469 "config": [ 00:23:53.469 { 00:23:53.469 "method": "accel_set_options", 00:23:53.469 "params": { 00:23:53.469 "small_cache_size": 128, 00:23:53.469 "large_cache_size": 16, 00:23:53.469 "task_count": 2048, 00:23:53.469 "sequence_count": 2048, 00:23:53.469 "buf_count": 2048 00:23:53.469 } 00:23:53.469 } 00:23:53.469 ] 00:23:53.469 }, 00:23:53.469 { 00:23:53.469 "subsystem": "bdev", 00:23:53.469 "config": [ 00:23:53.469 { 00:23:53.469 "method": "bdev_set_options", 00:23:53.469 "params": { 00:23:53.469 "bdev_io_pool_size": 65535, 00:23:53.469 "bdev_io_cache_size": 256, 00:23:53.469 "bdev_auto_examine": true, 00:23:53.469 "iobuf_small_cache_size": 128, 00:23:53.469 "iobuf_large_cache_size": 16 00:23:53.469 } 00:23:53.469 }, 00:23:53.470 { 00:23:53.470 "method": "bdev_raid_set_options", 00:23:53.470 "params": { 00:23:53.470 "process_window_size_kb": 1024, 00:23:53.470 "process_max_bandwidth_mb_sec": 0 00:23:53.470 } 00:23:53.470 }, 00:23:53.470 { 00:23:53.470 "method": "bdev_iscsi_set_options", 00:23:53.470 "params": { 00:23:53.470 "timeout_sec": 30 00:23:53.470 } 00:23:53.470 }, 00:23:53.470 { 00:23:53.470 "method": "bdev_nvme_set_options", 00:23:53.470 "params": { 00:23:53.470 "action_on_timeout": "none", 00:23:53.470 "timeout_us": 0, 00:23:53.470 "timeout_admin_us": 0, 00:23:53.470 "keep_alive_timeout_ms": 10000, 00:23:53.470 "arbitration_burst": 0, 00:23:53.470 "low_priority_weight": 0, 00:23:53.470 "medium_priority_weight": 0, 00:23:53.470 "high_priority_weight": 0, 00:23:53.470 "nvme_adminq_poll_period_us": 10000, 00:23:53.470 "nvme_ioq_poll_period_us": 0, 00:23:53.470 "io_queue_requests": 512, 00:23:53.470 "delay_cmd_submit": true, 00:23:53.470 "transport_retry_count": 4, 00:23:53.470 "bdev_retry_count": 3, 00:23:53.470 "transport_ack_timeout": 0, 00:23:53.470 "ctrlr_loss_timeout_sec": 0, 00:23:53.470 "reconnect_delay_sec": 0, 00:23:53.470 "fast_io_fail_timeout_sec": 0, 00:23:53.470 "disable_auto_failback": false, 00:23:53.470 "generate_uuids": false, 00:23:53.470 "transport_tos": 0, 00:23:53.470 "nvme_error_stat": false, 00:23:53.470 "rdma_srq_size": 0, 00:23:53.470 "io_path_stat": false, 00:23:53.470 "allow_accel_sequence": false, 00:23:53.470 "rdma_max_cq_size": 0, 00:23:53.470 "rdma_cm_event_timeout_ms": 0, 00:23:53.470 "dhchap_digests": [ 00:23:53.470 "sha256", 00:23:53.470 "sha384", 00:23:53.470 "sha512" 00:23:53.470 ], 00:23:53.470 "dhchap_dhgroups": [ 00:23:53.470 "null", 00:23:53.470 "ffdhe2048", 00:23:53.470 "ffdhe3072", 00:23:53.470 "ffdhe4096", 00:23:53.470 "ffdhe6144", 00:23:53.470 "ffdhe8192" 00:23:53.470 ] 00:23:53.470 } 00:23:53.470 }, 00:23:53.470 { 00:23:53.470 "method": "bdev_nvme_attach_controller", 00:23:53.470 "params": { 00:23:53.470 "name": "TLSTEST", 00:23:53.470 "trtype": "TCP", 00:23:53.470 "adrfam": "IPv4", 00:23:53.470 "traddr": "10.0.0.2", 00:23:53.470 "trsvcid": "4420", 00:23:53.470 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:23:53.470 "prchk_reftag": false, 00:23:53.470 "prchk_guard": false, 00:23:53.470 "ctrlr_loss_timeout_sec": 0, 00:23:53.470 "reconnect_delay_sec": 0, 00:23:53.470 "fast_io_fail_timeout_sec": 0, 00:23:53.470 "psk": "key0", 00:23:53.470 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:23:53.470 "hdgst": false, 00:23:53.470 "ddgst": false, 00:23:53.470 "multipath": "multipath" 00:23:53.470 } 00:23:53.470 }, 00:23:53.470 { 00:23:53.470 "method": "bdev_nvme_set_hotplug", 00:23:53.470 "params": { 00:23:53.470 "period_us": 100000, 00:23:53.470 "enable": false 00:23:53.470 } 00:23:53.470 }, 00:23:53.470 { 00:23:53.470 "method": "bdev_wait_for_examine" 00:23:53.470 } 00:23:53.470 ] 00:23:53.470 }, 00:23:53.470 { 00:23:53.470 "subsystem": "nbd", 00:23:53.470 "config": [] 00:23:53.470 } 00:23:53.470 ] 00:23:53.470 }' 00:23:53.470 [2024-11-06 08:59:43.438960] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:23:53.470 [2024-11-06 08:59:43.439014] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1917145 ] 00:23:53.470 [2024-11-06 08:59:43.497593] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:53.470 [2024-11-06 08:59:43.527568] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:23:53.731 [2024-11-06 08:59:43.662561] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:23:54.303 08:59:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:23:54.303 08:59:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:23:54.303 08:59:44 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@213 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:23:54.303 Running I/O for 10 seconds... 00:23:56.632 5231.00 IOPS, 20.43 MiB/s [2024-11-06T07:59:47.689Z] 5474.50 IOPS, 21.38 MiB/s [2024-11-06T07:59:48.632Z] 5813.33 IOPS, 22.71 MiB/s [2024-11-06T07:59:49.573Z] 5562.25 IOPS, 21.73 MiB/s [2024-11-06T07:59:50.518Z] 5515.20 IOPS, 21.54 MiB/s [2024-11-06T07:59:51.457Z] 5537.83 IOPS, 21.63 MiB/s [2024-11-06T07:59:52.397Z] 5677.00 IOPS, 22.18 MiB/s [2024-11-06T07:59:53.782Z] 5620.75 IOPS, 21.96 MiB/s [2024-11-06T07:59:54.353Z] 5656.67 IOPS, 22.10 MiB/s [2024-11-06T07:59:54.614Z] 5625.70 IOPS, 21.98 MiB/s 00:24:04.501 Latency(us) 00:24:04.501 [2024-11-06T07:59:54.614Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:04.501 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:24:04.501 Verification LBA range: start 0x0 length 0x2000 00:24:04.501 TLSTESTn1 : 10.05 5610.31 21.92 0.00 0.00 22747.94 4696.75 45875.20 00:24:04.501 [2024-11-06T07:59:54.614Z] =================================================================================================================== 00:24:04.501 [2024-11-06T07:59:54.614Z] Total : 5610.31 21.92 0.00 0.00 22747.94 4696.75 45875.20 00:24:04.501 { 00:24:04.501 "results": [ 00:24:04.501 { 00:24:04.501 "job": "TLSTESTn1", 00:24:04.501 "core_mask": "0x4", 00:24:04.501 "workload": "verify", 00:24:04.501 "status": "finished", 00:24:04.501 "verify_range": { 00:24:04.501 "start": 0, 00:24:04.501 "length": 8192 00:24:04.501 }, 00:24:04.501 "queue_depth": 128, 00:24:04.501 "io_size": 4096, 00:24:04.501 "runtime": 10.050254, 00:24:04.501 "iops": 5610.30596838647, 00:24:04.501 "mibps": 21.91525768900965, 00:24:04.501 "io_failed": 0, 00:24:04.501 "io_timeout": 0, 00:24:04.501 "avg_latency_us": 22747.9408287074, 00:24:04.501 "min_latency_us": 4696.746666666667, 00:24:04.501 "max_latency_us": 45875.2 00:24:04.501 } 00:24:04.501 ], 00:24:04.501 "core_count": 1 00:24:04.501 } 00:24:04.501 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@215 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:24:04.501 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@216 -- # killprocess 1917145 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1917145 ']' 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1917145 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1917145 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1917145' 00:24:04.502 killing process with pid 1917145 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1917145 00:24:04.502 Received shutdown signal, test time was about 10.000000 seconds 00:24:04.502 00:24:04.502 Latency(us) 00:24:04.502 [2024-11-06T07:59:54.615Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:04.502 [2024-11-06T07:59:54.615Z] =================================================================================================================== 00:24:04.502 [2024-11-06T07:59:54.615Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1917145 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@217 -- # killprocess 1916800 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1916800 ']' 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1916800 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:04.502 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1916800 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1916800' 00:24:04.763 killing process with pid 1916800 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1916800 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1916800 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@220 -- # nvmfappstart 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=1919170 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 1919170 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1919170 ']' 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:04.763 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:04.763 08:59:54 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:04.763 [2024-11-06 08:59:54.823649] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:24:04.763 [2024-11-06 08:59:54.823704] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:05.024 [2024-11-06 08:59:54.901979] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:05.024 [2024-11-06 08:59:54.936967] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:05.024 [2024-11-06 08:59:54.937000] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:05.024 [2024-11-06 08:59:54.937008] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:05.024 [2024-11-06 08:59:54.937015] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:05.024 [2024-11-06 08:59:54.937020] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:05.024 [2024-11-06 08:59:54.937574] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:05.595 08:59:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:05.595 08:59:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:24:05.595 08:59:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:24:05.595 08:59:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:24:05.595 08:59:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:05.595 08:59:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:05.595 08:59:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@221 -- # setup_nvmf_tgt /tmp/tmp.mY106GdROb 00:24:05.595 08:59:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@50 -- # local key=/tmp/tmp.mY106GdROb 00:24:05.595 08:59:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:24:05.855 [2024-11-06 08:59:55.807277] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:05.855 08:59:55 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:24:06.116 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:24:06.116 [2024-11-06 08:59:56.168174] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:24:06.116 [2024-11-06 08:59:56.168392] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:06.116 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:24:06.376 malloc0 00:24:06.376 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:24:06.637 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py keyring_file_add_key key0 /tmp/tmp.mY106GdROb 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk key0 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@222 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@224 -- # bdevperf_pid=1919717 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@226 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@227 -- # waitforlisten 1919717 /var/tmp/bdevperf.sock 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1919717 ']' 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:24:06.898 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:06.898 08:59:56 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:06.898 [2024-11-06 08:59:56.974867] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:24:06.898 [2024-11-06 08:59:56.974919] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1919717 ] 00:24:07.158 [2024-11-06 08:59:57.059111] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:07.158 [2024-11-06 08:59:57.088837] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:07.158 08:59:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:07.158 08:59:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:24:07.158 08:59:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@229 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.mY106GdROb 00:24:07.419 08:59:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@230 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:24:07.419 [2024-11-06 08:59:57.467813] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:24:07.679 nvme0n1 00:24:07.679 08:59:57 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@234 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:24:07.679 Running I/O for 1 seconds... 00:24:08.621 5172.00 IOPS, 20.20 MiB/s 00:24:08.621 Latency(us) 00:24:08.621 [2024-11-06T07:59:58.734Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:08.621 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:24:08.621 Verification LBA range: start 0x0 length 0x2000 00:24:08.621 nvme0n1 : 1.02 5216.61 20.38 0.00 0.00 24371.74 6990.51 53084.16 00:24:08.621 [2024-11-06T07:59:58.734Z] =================================================================================================================== 00:24:08.621 [2024-11-06T07:59:58.734Z] Total : 5216.61 20.38 0.00 0.00 24371.74 6990.51 53084.16 00:24:08.621 { 00:24:08.621 "results": [ 00:24:08.621 { 00:24:08.621 "job": "nvme0n1", 00:24:08.621 "core_mask": "0x2", 00:24:08.621 "workload": "verify", 00:24:08.621 "status": "finished", 00:24:08.621 "verify_range": { 00:24:08.621 "start": 0, 00:24:08.621 "length": 8192 00:24:08.621 }, 00:24:08.621 "queue_depth": 128, 00:24:08.621 "io_size": 4096, 00:24:08.621 "runtime": 1.016177, 00:24:08.621 "iops": 5216.610885701999, 00:24:08.621 "mibps": 20.377386272273434, 00:24:08.621 "io_failed": 0, 00:24:08.621 "io_timeout": 0, 00:24:08.621 "avg_latency_us": 24371.743454694082, 00:24:08.621 "min_latency_us": 6990.506666666667, 00:24:08.621 "max_latency_us": 53084.16 00:24:08.621 } 00:24:08.621 ], 00:24:08.621 "core_count": 1 00:24:08.621 } 00:24:08.621 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@236 -- # killprocess 1919717 00:24:08.621 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1919717 ']' 00:24:08.621 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1919717 00:24:08.621 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:24:08.621 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:08.621 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1919717 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1919717' 00:24:08.882 killing process with pid 1919717 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1919717 00:24:08.882 Received shutdown signal, test time was about 1.000000 seconds 00:24:08.882 00:24:08.882 Latency(us) 00:24:08.882 [2024-11-06T07:59:58.995Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:08.882 [2024-11-06T07:59:58.995Z] =================================================================================================================== 00:24:08.882 [2024-11-06T07:59:58.995Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1919717 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@237 -- # killprocess 1919170 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1919170 ']' 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1919170 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1919170 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1919170' 00:24:08.882 killing process with pid 1919170 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1919170 00:24:08.882 08:59:58 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1919170 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@242 -- # nvmfappstart 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=1920202 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 1920202 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1920202 ']' 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:09.143 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:09.143 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:09.143 [2024-11-06 08:59:59.088082] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:24:09.143 [2024-11-06 08:59:59.088135] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:09.143 [2024-11-06 08:59:59.164648] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:09.143 [2024-11-06 08:59:59.198295] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:09.143 [2024-11-06 08:59:59.198331] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:09.143 [2024-11-06 08:59:59.198341] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:09.143 [2024-11-06 08:59:59.198348] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:09.143 [2024-11-06 08:59:59.198353] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:09.143 [2024-11-06 08:59:59.198930] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@243 -- # rpc_cmd 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:10.091 [2024-11-06 08:59:59.932176] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:10.091 malloc0 00:24:10.091 [2024-11-06 08:59:59.958873] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:24:10.091 [2024-11-06 08:59:59.959092] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@256 -- # bdevperf_pid=1920244 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@258 -- # waitforlisten 1920244 /var/tmp/bdevperf.sock 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@254 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1920244 ']' 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:24:10.091 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:10.091 08:59:59 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:10.091 [2024-11-06 09:00:00.038670] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:24:10.091 [2024-11-06 09:00:00.038721] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1920244 ] 00:24:10.091 [2024-11-06 09:00:00.123619] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:10.091 [2024-11-06 09:00:00.154048] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:11.032 09:00:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:11.032 09:00:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:24:11.032 09:00:00 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@259 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.mY106GdROb 00:24:11.032 09:00:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@260 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:24:11.294 [2024-11-06 09:00:01.166675] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:24:11.294 nvme0n1 00:24:11.294 09:00:01 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@264 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:24:11.294 Running I/O for 1 seconds... 00:24:12.678 4139.00 IOPS, 16.17 MiB/s 00:24:12.678 Latency(us) 00:24:12.678 [2024-11-06T08:00:02.791Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:12.678 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:24:12.678 Verification LBA range: start 0x0 length 0x2000 00:24:12.678 nvme0n1 : 1.08 3953.10 15.44 0.00 0.00 31612.38 5789.01 89565.87 00:24:12.678 [2024-11-06T08:00:02.791Z] =================================================================================================================== 00:24:12.678 [2024-11-06T08:00:02.791Z] Total : 3953.10 15.44 0.00 0.00 31612.38 5789.01 89565.87 00:24:12.678 { 00:24:12.678 "results": [ 00:24:12.678 { 00:24:12.678 "job": "nvme0n1", 00:24:12.678 "core_mask": "0x2", 00:24:12.678 "workload": "verify", 00:24:12.678 "status": "finished", 00:24:12.678 "verify_range": { 00:24:12.678 "start": 0, 00:24:12.678 "length": 8192 00:24:12.678 }, 00:24:12.678 "queue_depth": 128, 00:24:12.678 "io_size": 4096, 00:24:12.678 "runtime": 1.079406, 00:24:12.678 "iops": 3953.1001309979747, 00:24:12.678 "mibps": 15.441797386710839, 00:24:12.678 "io_failed": 0, 00:24:12.678 "io_timeout": 0, 00:24:12.678 "avg_latency_us": 31612.378282946647, 00:24:12.678 "min_latency_us": 5789.013333333333, 00:24:12.678 "max_latency_us": 89565.86666666667 00:24:12.678 } 00:24:12.678 ], 00:24:12.678 "core_count": 1 00:24:12.678 } 00:24:12.678 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@267 -- # rpc_cmd save_config 00:24:12.678 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:12.678 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:12.678 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:12.678 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@267 -- # tgtcfg='{ 00:24:12.678 "subsystems": [ 00:24:12.678 { 00:24:12.678 "subsystem": "keyring", 00:24:12.678 "config": [ 00:24:12.678 { 00:24:12.678 "method": "keyring_file_add_key", 00:24:12.678 "params": { 00:24:12.678 "name": "key0", 00:24:12.678 "path": "/tmp/tmp.mY106GdROb" 00:24:12.678 } 00:24:12.678 } 00:24:12.678 ] 00:24:12.678 }, 00:24:12.678 { 00:24:12.678 "subsystem": "iobuf", 00:24:12.678 "config": [ 00:24:12.678 { 00:24:12.678 "method": "iobuf_set_options", 00:24:12.678 "params": { 00:24:12.678 "small_pool_count": 8192, 00:24:12.678 "large_pool_count": 1024, 00:24:12.678 "small_bufsize": 8192, 00:24:12.678 "large_bufsize": 135168, 00:24:12.678 "enable_numa": false 00:24:12.678 } 00:24:12.678 } 00:24:12.678 ] 00:24:12.678 }, 00:24:12.678 { 00:24:12.678 "subsystem": "sock", 00:24:12.678 "config": [ 00:24:12.678 { 00:24:12.678 "method": "sock_set_default_impl", 00:24:12.678 "params": { 00:24:12.678 "impl_name": "posix" 00:24:12.678 } 00:24:12.678 }, 00:24:12.678 { 00:24:12.678 "method": "sock_impl_set_options", 00:24:12.678 "params": { 00:24:12.678 "impl_name": "ssl", 00:24:12.678 "recv_buf_size": 4096, 00:24:12.678 "send_buf_size": 4096, 00:24:12.678 "enable_recv_pipe": true, 00:24:12.678 "enable_quickack": false, 00:24:12.678 "enable_placement_id": 0, 00:24:12.678 "enable_zerocopy_send_server": true, 00:24:12.678 "enable_zerocopy_send_client": false, 00:24:12.678 "zerocopy_threshold": 0, 00:24:12.678 "tls_version": 0, 00:24:12.678 "enable_ktls": false 00:24:12.678 } 00:24:12.678 }, 00:24:12.678 { 00:24:12.678 "method": "sock_impl_set_options", 00:24:12.678 "params": { 00:24:12.678 "impl_name": "posix", 00:24:12.678 "recv_buf_size": 2097152, 00:24:12.678 "send_buf_size": 2097152, 00:24:12.678 "enable_recv_pipe": true, 00:24:12.678 "enable_quickack": false, 00:24:12.678 "enable_placement_id": 0, 00:24:12.678 "enable_zerocopy_send_server": true, 00:24:12.678 "enable_zerocopy_send_client": false, 00:24:12.678 "zerocopy_threshold": 0, 00:24:12.678 "tls_version": 0, 00:24:12.678 "enable_ktls": false 00:24:12.678 } 00:24:12.678 } 00:24:12.678 ] 00:24:12.678 }, 00:24:12.678 { 00:24:12.678 "subsystem": "vmd", 00:24:12.678 "config": [] 00:24:12.678 }, 00:24:12.678 { 00:24:12.678 "subsystem": "accel", 00:24:12.678 "config": [ 00:24:12.678 { 00:24:12.678 "method": "accel_set_options", 00:24:12.678 "params": { 00:24:12.678 "small_cache_size": 128, 00:24:12.678 "large_cache_size": 16, 00:24:12.678 "task_count": 2048, 00:24:12.678 "sequence_count": 2048, 00:24:12.678 "buf_count": 2048 00:24:12.678 } 00:24:12.678 } 00:24:12.678 ] 00:24:12.678 }, 00:24:12.678 { 00:24:12.678 "subsystem": "bdev", 00:24:12.678 "config": [ 00:24:12.678 { 00:24:12.678 "method": "bdev_set_options", 00:24:12.678 "params": { 00:24:12.678 "bdev_io_pool_size": 65535, 00:24:12.678 "bdev_io_cache_size": 256, 00:24:12.678 "bdev_auto_examine": true, 00:24:12.678 "iobuf_small_cache_size": 128, 00:24:12.678 "iobuf_large_cache_size": 16 00:24:12.678 } 00:24:12.678 }, 00:24:12.678 { 00:24:12.678 "method": "bdev_raid_set_options", 00:24:12.678 "params": { 00:24:12.678 "process_window_size_kb": 1024, 00:24:12.678 "process_max_bandwidth_mb_sec": 0 00:24:12.678 } 00:24:12.678 }, 00:24:12.678 { 00:24:12.678 "method": "bdev_iscsi_set_options", 00:24:12.678 "params": { 00:24:12.678 "timeout_sec": 30 00:24:12.678 } 00:24:12.678 }, 00:24:12.678 { 00:24:12.678 "method": "bdev_nvme_set_options", 00:24:12.678 "params": { 00:24:12.678 "action_on_timeout": "none", 00:24:12.678 "timeout_us": 0, 00:24:12.678 "timeout_admin_us": 0, 00:24:12.679 "keep_alive_timeout_ms": 10000, 00:24:12.679 "arbitration_burst": 0, 00:24:12.679 "low_priority_weight": 0, 00:24:12.679 "medium_priority_weight": 0, 00:24:12.679 "high_priority_weight": 0, 00:24:12.679 "nvme_adminq_poll_period_us": 10000, 00:24:12.679 "nvme_ioq_poll_period_us": 0, 00:24:12.679 "io_queue_requests": 0, 00:24:12.679 "delay_cmd_submit": true, 00:24:12.679 "transport_retry_count": 4, 00:24:12.679 "bdev_retry_count": 3, 00:24:12.679 "transport_ack_timeout": 0, 00:24:12.679 "ctrlr_loss_timeout_sec": 0, 00:24:12.679 "reconnect_delay_sec": 0, 00:24:12.679 "fast_io_fail_timeout_sec": 0, 00:24:12.679 "disable_auto_failback": false, 00:24:12.679 "generate_uuids": false, 00:24:12.679 "transport_tos": 0, 00:24:12.679 "nvme_error_stat": false, 00:24:12.679 "rdma_srq_size": 0, 00:24:12.679 "io_path_stat": false, 00:24:12.679 "allow_accel_sequence": false, 00:24:12.679 "rdma_max_cq_size": 0, 00:24:12.679 "rdma_cm_event_timeout_ms": 0, 00:24:12.679 "dhchap_digests": [ 00:24:12.679 "sha256", 00:24:12.679 "sha384", 00:24:12.679 "sha512" 00:24:12.679 ], 00:24:12.679 "dhchap_dhgroups": [ 00:24:12.679 "null", 00:24:12.679 "ffdhe2048", 00:24:12.679 "ffdhe3072", 00:24:12.679 "ffdhe4096", 00:24:12.679 "ffdhe6144", 00:24:12.679 "ffdhe8192" 00:24:12.679 ] 00:24:12.679 } 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "method": "bdev_nvme_set_hotplug", 00:24:12.679 "params": { 00:24:12.679 "period_us": 100000, 00:24:12.679 "enable": false 00:24:12.679 } 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "method": "bdev_malloc_create", 00:24:12.679 "params": { 00:24:12.679 "name": "malloc0", 00:24:12.679 "num_blocks": 8192, 00:24:12.679 "block_size": 4096, 00:24:12.679 "physical_block_size": 4096, 00:24:12.679 "uuid": "22627714-e34c-4838-9d31-5341266af307", 00:24:12.679 "optimal_io_boundary": 0, 00:24:12.679 "md_size": 0, 00:24:12.679 "dif_type": 0, 00:24:12.679 "dif_is_head_of_md": false, 00:24:12.679 "dif_pi_format": 0 00:24:12.679 } 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "method": "bdev_wait_for_examine" 00:24:12.679 } 00:24:12.679 ] 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "subsystem": "nbd", 00:24:12.679 "config": [] 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "subsystem": "scheduler", 00:24:12.679 "config": [ 00:24:12.679 { 00:24:12.679 "method": "framework_set_scheduler", 00:24:12.679 "params": { 00:24:12.679 "name": "static" 00:24:12.679 } 00:24:12.679 } 00:24:12.679 ] 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "subsystem": "nvmf", 00:24:12.679 "config": [ 00:24:12.679 { 00:24:12.679 "method": "nvmf_set_config", 00:24:12.679 "params": { 00:24:12.679 "discovery_filter": "match_any", 00:24:12.679 "admin_cmd_passthru": { 00:24:12.679 "identify_ctrlr": false 00:24:12.679 }, 00:24:12.679 "dhchap_digests": [ 00:24:12.679 "sha256", 00:24:12.679 "sha384", 00:24:12.679 "sha512" 00:24:12.679 ], 00:24:12.679 "dhchap_dhgroups": [ 00:24:12.679 "null", 00:24:12.679 "ffdhe2048", 00:24:12.679 "ffdhe3072", 00:24:12.679 "ffdhe4096", 00:24:12.679 "ffdhe6144", 00:24:12.679 "ffdhe8192" 00:24:12.679 ] 00:24:12.679 } 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "method": "nvmf_set_max_subsystems", 00:24:12.679 "params": { 00:24:12.679 "max_subsystems": 1024 00:24:12.679 } 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "method": "nvmf_set_crdt", 00:24:12.679 "params": { 00:24:12.679 "crdt1": 0, 00:24:12.679 "crdt2": 0, 00:24:12.679 "crdt3": 0 00:24:12.679 } 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "method": "nvmf_create_transport", 00:24:12.679 "params": { 00:24:12.679 "trtype": "TCP", 00:24:12.679 "max_queue_depth": 128, 00:24:12.679 "max_io_qpairs_per_ctrlr": 127, 00:24:12.679 "in_capsule_data_size": 4096, 00:24:12.679 "max_io_size": 131072, 00:24:12.679 "io_unit_size": 131072, 00:24:12.679 "max_aq_depth": 128, 00:24:12.679 "num_shared_buffers": 511, 00:24:12.679 "buf_cache_size": 4294967295, 00:24:12.679 "dif_insert_or_strip": false, 00:24:12.679 "zcopy": false, 00:24:12.679 "c2h_success": false, 00:24:12.679 "sock_priority": 0, 00:24:12.679 "abort_timeout_sec": 1, 00:24:12.679 "ack_timeout": 0, 00:24:12.679 "data_wr_pool_size": 0 00:24:12.679 } 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "method": "nvmf_create_subsystem", 00:24:12.679 "params": { 00:24:12.679 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:12.679 "allow_any_host": false, 00:24:12.679 "serial_number": "00000000000000000000", 00:24:12.679 "model_number": "SPDK bdev Controller", 00:24:12.679 "max_namespaces": 32, 00:24:12.679 "min_cntlid": 1, 00:24:12.679 "max_cntlid": 65519, 00:24:12.679 "ana_reporting": false 00:24:12.679 } 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "method": "nvmf_subsystem_add_host", 00:24:12.679 "params": { 00:24:12.679 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:12.679 "host": "nqn.2016-06.io.spdk:host1", 00:24:12.679 "psk": "key0" 00:24:12.679 } 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "method": "nvmf_subsystem_add_ns", 00:24:12.679 "params": { 00:24:12.679 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:12.679 "namespace": { 00:24:12.679 "nsid": 1, 00:24:12.679 "bdev_name": "malloc0", 00:24:12.679 "nguid": "22627714E34C48389D315341266AF307", 00:24:12.679 "uuid": "22627714-e34c-4838-9d31-5341266af307", 00:24:12.679 "no_auto_visible": false 00:24:12.679 } 00:24:12.679 } 00:24:12.679 }, 00:24:12.679 { 00:24:12.679 "method": "nvmf_subsystem_add_listener", 00:24:12.679 "params": { 00:24:12.679 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:12.679 "listen_address": { 00:24:12.679 "trtype": "TCP", 00:24:12.679 "adrfam": "IPv4", 00:24:12.679 "traddr": "10.0.0.2", 00:24:12.679 "trsvcid": "4420" 00:24:12.679 }, 00:24:12.679 "secure_channel": false, 00:24:12.679 "sock_impl": "ssl" 00:24:12.679 } 00:24:12.679 } 00:24:12.679 ] 00:24:12.679 } 00:24:12.679 ] 00:24:12.679 }' 00:24:12.679 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@268 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:24:12.940 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@268 -- # bperfcfg='{ 00:24:12.940 "subsystems": [ 00:24:12.940 { 00:24:12.940 "subsystem": "keyring", 00:24:12.940 "config": [ 00:24:12.940 { 00:24:12.940 "method": "keyring_file_add_key", 00:24:12.940 "params": { 00:24:12.940 "name": "key0", 00:24:12.940 "path": "/tmp/tmp.mY106GdROb" 00:24:12.940 } 00:24:12.940 } 00:24:12.940 ] 00:24:12.940 }, 00:24:12.940 { 00:24:12.940 "subsystem": "iobuf", 00:24:12.940 "config": [ 00:24:12.940 { 00:24:12.940 "method": "iobuf_set_options", 00:24:12.940 "params": { 00:24:12.940 "small_pool_count": 8192, 00:24:12.940 "large_pool_count": 1024, 00:24:12.940 "small_bufsize": 8192, 00:24:12.940 "large_bufsize": 135168, 00:24:12.940 "enable_numa": false 00:24:12.940 } 00:24:12.940 } 00:24:12.940 ] 00:24:12.940 }, 00:24:12.940 { 00:24:12.940 "subsystem": "sock", 00:24:12.940 "config": [ 00:24:12.940 { 00:24:12.940 "method": "sock_set_default_impl", 00:24:12.940 "params": { 00:24:12.940 "impl_name": "posix" 00:24:12.940 } 00:24:12.940 }, 00:24:12.940 { 00:24:12.940 "method": "sock_impl_set_options", 00:24:12.940 "params": { 00:24:12.940 "impl_name": "ssl", 00:24:12.940 "recv_buf_size": 4096, 00:24:12.940 "send_buf_size": 4096, 00:24:12.940 "enable_recv_pipe": true, 00:24:12.940 "enable_quickack": false, 00:24:12.940 "enable_placement_id": 0, 00:24:12.940 "enable_zerocopy_send_server": true, 00:24:12.940 "enable_zerocopy_send_client": false, 00:24:12.940 "zerocopy_threshold": 0, 00:24:12.940 "tls_version": 0, 00:24:12.940 "enable_ktls": false 00:24:12.940 } 00:24:12.940 }, 00:24:12.940 { 00:24:12.940 "method": "sock_impl_set_options", 00:24:12.940 "params": { 00:24:12.940 "impl_name": "posix", 00:24:12.940 "recv_buf_size": 2097152, 00:24:12.940 "send_buf_size": 2097152, 00:24:12.940 "enable_recv_pipe": true, 00:24:12.940 "enable_quickack": false, 00:24:12.940 "enable_placement_id": 0, 00:24:12.940 "enable_zerocopy_send_server": true, 00:24:12.940 "enable_zerocopy_send_client": false, 00:24:12.940 "zerocopy_threshold": 0, 00:24:12.940 "tls_version": 0, 00:24:12.940 "enable_ktls": false 00:24:12.940 } 00:24:12.940 } 00:24:12.940 ] 00:24:12.940 }, 00:24:12.940 { 00:24:12.940 "subsystem": "vmd", 00:24:12.940 "config": [] 00:24:12.940 }, 00:24:12.940 { 00:24:12.940 "subsystem": "accel", 00:24:12.940 "config": [ 00:24:12.940 { 00:24:12.940 "method": "accel_set_options", 00:24:12.940 "params": { 00:24:12.940 "small_cache_size": 128, 00:24:12.940 "large_cache_size": 16, 00:24:12.940 "task_count": 2048, 00:24:12.940 "sequence_count": 2048, 00:24:12.940 "buf_count": 2048 00:24:12.940 } 00:24:12.940 } 00:24:12.940 ] 00:24:12.940 }, 00:24:12.940 { 00:24:12.940 "subsystem": "bdev", 00:24:12.940 "config": [ 00:24:12.940 { 00:24:12.940 "method": "bdev_set_options", 00:24:12.940 "params": { 00:24:12.940 "bdev_io_pool_size": 65535, 00:24:12.940 "bdev_io_cache_size": 256, 00:24:12.940 "bdev_auto_examine": true, 00:24:12.940 "iobuf_small_cache_size": 128, 00:24:12.940 "iobuf_large_cache_size": 16 00:24:12.940 } 00:24:12.940 }, 00:24:12.940 { 00:24:12.940 "method": "bdev_raid_set_options", 00:24:12.940 "params": { 00:24:12.940 "process_window_size_kb": 1024, 00:24:12.940 "process_max_bandwidth_mb_sec": 0 00:24:12.940 } 00:24:12.940 }, 00:24:12.940 { 00:24:12.940 "method": "bdev_iscsi_set_options", 00:24:12.940 "params": { 00:24:12.940 "timeout_sec": 30 00:24:12.940 } 00:24:12.940 }, 00:24:12.940 { 00:24:12.940 "method": "bdev_nvme_set_options", 00:24:12.940 "params": { 00:24:12.940 "action_on_timeout": "none", 00:24:12.940 "timeout_us": 0, 00:24:12.940 "timeout_admin_us": 0, 00:24:12.941 "keep_alive_timeout_ms": 10000, 00:24:12.941 "arbitration_burst": 0, 00:24:12.941 "low_priority_weight": 0, 00:24:12.941 "medium_priority_weight": 0, 00:24:12.941 "high_priority_weight": 0, 00:24:12.941 "nvme_adminq_poll_period_us": 10000, 00:24:12.941 "nvme_ioq_poll_period_us": 0, 00:24:12.941 "io_queue_requests": 512, 00:24:12.941 "delay_cmd_submit": true, 00:24:12.941 "transport_retry_count": 4, 00:24:12.941 "bdev_retry_count": 3, 00:24:12.941 "transport_ack_timeout": 0, 00:24:12.941 "ctrlr_loss_timeout_sec": 0, 00:24:12.941 "reconnect_delay_sec": 0, 00:24:12.941 "fast_io_fail_timeout_sec": 0, 00:24:12.941 "disable_auto_failback": false, 00:24:12.941 "generate_uuids": false, 00:24:12.941 "transport_tos": 0, 00:24:12.941 "nvme_error_stat": false, 00:24:12.941 "rdma_srq_size": 0, 00:24:12.941 "io_path_stat": false, 00:24:12.941 "allow_accel_sequence": false, 00:24:12.941 "rdma_max_cq_size": 0, 00:24:12.941 "rdma_cm_event_timeout_ms": 0, 00:24:12.941 "dhchap_digests": [ 00:24:12.941 "sha256", 00:24:12.941 "sha384", 00:24:12.941 "sha512" 00:24:12.941 ], 00:24:12.941 "dhchap_dhgroups": [ 00:24:12.941 "null", 00:24:12.941 "ffdhe2048", 00:24:12.941 "ffdhe3072", 00:24:12.941 "ffdhe4096", 00:24:12.941 "ffdhe6144", 00:24:12.941 "ffdhe8192" 00:24:12.941 ] 00:24:12.941 } 00:24:12.941 }, 00:24:12.941 { 00:24:12.941 "method": "bdev_nvme_attach_controller", 00:24:12.941 "params": { 00:24:12.941 "name": "nvme0", 00:24:12.941 "trtype": "TCP", 00:24:12.941 "adrfam": "IPv4", 00:24:12.941 "traddr": "10.0.0.2", 00:24:12.941 "trsvcid": "4420", 00:24:12.941 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:24:12.941 "prchk_reftag": false, 00:24:12.941 "prchk_guard": false, 00:24:12.941 "ctrlr_loss_timeout_sec": 0, 00:24:12.941 "reconnect_delay_sec": 0, 00:24:12.941 "fast_io_fail_timeout_sec": 0, 00:24:12.941 "psk": "key0", 00:24:12.941 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:24:12.941 "hdgst": false, 00:24:12.941 "ddgst": false, 00:24:12.941 "multipath": "multipath" 00:24:12.941 } 00:24:12.941 }, 00:24:12.941 { 00:24:12.941 "method": "bdev_nvme_set_hotplug", 00:24:12.941 "params": { 00:24:12.941 "period_us": 100000, 00:24:12.941 "enable": false 00:24:12.941 } 00:24:12.941 }, 00:24:12.941 { 00:24:12.941 "method": "bdev_enable_histogram", 00:24:12.941 "params": { 00:24:12.941 "name": "nvme0n1", 00:24:12.941 "enable": true 00:24:12.941 } 00:24:12.941 }, 00:24:12.941 { 00:24:12.941 "method": "bdev_wait_for_examine" 00:24:12.941 } 00:24:12.941 ] 00:24:12.941 }, 00:24:12.941 { 00:24:12.941 "subsystem": "nbd", 00:24:12.941 "config": [] 00:24:12.941 } 00:24:12.941 ] 00:24:12.941 }' 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@270 -- # killprocess 1920244 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1920244 ']' 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1920244 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1920244 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1920244' 00:24:12.941 killing process with pid 1920244 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1920244 00:24:12.941 Received shutdown signal, test time was about 1.000000 seconds 00:24:12.941 00:24:12.941 Latency(us) 00:24:12.941 [2024-11-06T08:00:03.054Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:12.941 [2024-11-06T08:00:03.054Z] =================================================================================================================== 00:24:12.941 [2024-11-06T08:00:03.054Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1920244 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@271 -- # killprocess 1920202 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1920202 ']' 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1920202 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:12.941 09:00:02 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1920202 00:24:12.941 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:12.941 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:12.941 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1920202' 00:24:12.941 killing process with pid 1920202 00:24:12.941 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1920202 00:24:12.941 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1920202 00:24:13.202 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@273 -- # nvmfappstart -c /dev/fd/62 00:24:13.202 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:24:13.202 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@724 -- # xtrace_disable 00:24:13.202 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@273 -- # echo '{ 00:24:13.202 "subsystems": [ 00:24:13.202 { 00:24:13.202 "subsystem": "keyring", 00:24:13.202 "config": [ 00:24:13.202 { 00:24:13.202 "method": "keyring_file_add_key", 00:24:13.202 "params": { 00:24:13.202 "name": "key0", 00:24:13.202 "path": "/tmp/tmp.mY106GdROb" 00:24:13.202 } 00:24:13.202 } 00:24:13.202 ] 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "subsystem": "iobuf", 00:24:13.202 "config": [ 00:24:13.202 { 00:24:13.202 "method": "iobuf_set_options", 00:24:13.202 "params": { 00:24:13.202 "small_pool_count": 8192, 00:24:13.202 "large_pool_count": 1024, 00:24:13.202 "small_bufsize": 8192, 00:24:13.202 "large_bufsize": 135168, 00:24:13.202 "enable_numa": false 00:24:13.202 } 00:24:13.202 } 00:24:13.202 ] 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "subsystem": "sock", 00:24:13.202 "config": [ 00:24:13.202 { 00:24:13.202 "method": "sock_set_default_impl", 00:24:13.202 "params": { 00:24:13.202 "impl_name": "posix" 00:24:13.202 } 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "method": "sock_impl_set_options", 00:24:13.202 "params": { 00:24:13.202 "impl_name": "ssl", 00:24:13.202 "recv_buf_size": 4096, 00:24:13.202 "send_buf_size": 4096, 00:24:13.202 "enable_recv_pipe": true, 00:24:13.202 "enable_quickack": false, 00:24:13.202 "enable_placement_id": 0, 00:24:13.202 "enable_zerocopy_send_server": true, 00:24:13.202 "enable_zerocopy_send_client": false, 00:24:13.202 "zerocopy_threshold": 0, 00:24:13.202 "tls_version": 0, 00:24:13.202 "enable_ktls": false 00:24:13.202 } 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "method": "sock_impl_set_options", 00:24:13.202 "params": { 00:24:13.202 "impl_name": "posix", 00:24:13.202 "recv_buf_size": 2097152, 00:24:13.202 "send_buf_size": 2097152, 00:24:13.202 "enable_recv_pipe": true, 00:24:13.202 "enable_quickack": false, 00:24:13.202 "enable_placement_id": 0, 00:24:13.202 "enable_zerocopy_send_server": true, 00:24:13.202 "enable_zerocopy_send_client": false, 00:24:13.202 "zerocopy_threshold": 0, 00:24:13.202 "tls_version": 0, 00:24:13.202 "enable_ktls": false 00:24:13.202 } 00:24:13.202 } 00:24:13.202 ] 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "subsystem": "vmd", 00:24:13.202 "config": [] 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "subsystem": "accel", 00:24:13.202 "config": [ 00:24:13.202 { 00:24:13.202 "method": "accel_set_options", 00:24:13.202 "params": { 00:24:13.202 "small_cache_size": 128, 00:24:13.202 "large_cache_size": 16, 00:24:13.202 "task_count": 2048, 00:24:13.202 "sequence_count": 2048, 00:24:13.202 "buf_count": 2048 00:24:13.202 } 00:24:13.202 } 00:24:13.202 ] 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "subsystem": "bdev", 00:24:13.202 "config": [ 00:24:13.202 { 00:24:13.202 "method": "bdev_set_options", 00:24:13.202 "params": { 00:24:13.202 "bdev_io_pool_size": 65535, 00:24:13.202 "bdev_io_cache_size": 256, 00:24:13.202 "bdev_auto_examine": true, 00:24:13.202 "iobuf_small_cache_size": 128, 00:24:13.202 "iobuf_large_cache_size": 16 00:24:13.202 } 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "method": "bdev_raid_set_options", 00:24:13.202 "params": { 00:24:13.202 "process_window_size_kb": 1024, 00:24:13.202 "process_max_bandwidth_mb_sec": 0 00:24:13.202 } 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "method": "bdev_iscsi_set_options", 00:24:13.202 "params": { 00:24:13.202 "timeout_sec": 30 00:24:13.202 } 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "method": "bdev_nvme_set_options", 00:24:13.202 "params": { 00:24:13.202 "action_on_timeout": "none", 00:24:13.202 "timeout_us": 0, 00:24:13.202 "timeout_admin_us": 0, 00:24:13.202 "keep_alive_timeout_ms": 10000, 00:24:13.202 "arbitration_burst": 0, 00:24:13.202 "low_priority_weight": 0, 00:24:13.202 "medium_priority_weight": 0, 00:24:13.202 "high_priority_weight": 0, 00:24:13.202 "nvme_adminq_poll_period_us": 10000, 00:24:13.202 "nvme_ioq_poll_period_us": 0, 00:24:13.202 "io_queue_requests": 0, 00:24:13.202 "delay_cmd_submit": true, 00:24:13.202 "transport_retry_count": 4, 00:24:13.202 "bdev_retry_count": 3, 00:24:13.202 "transport_ack_timeout": 0, 00:24:13.202 "ctrlr_loss_timeout_sec": 0, 00:24:13.202 "reconnect_delay_sec": 0, 00:24:13.202 "fast_io_fail_timeout_sec": 0, 00:24:13.202 "disable_auto_failback": false, 00:24:13.202 "generate_uuids": false, 00:24:13.202 "transport_tos": 0, 00:24:13.202 "nvme_error_stat": false, 00:24:13.202 "rdma_srq_size": 0, 00:24:13.202 "io_path_stat": false, 00:24:13.202 "allow_accel_sequence": false, 00:24:13.202 "rdma_max_cq_size": 0, 00:24:13.202 "rdma_cm_event_timeout_ms": 0, 00:24:13.202 "dhchap_digests": [ 00:24:13.202 "sha256", 00:24:13.202 "sha384", 00:24:13.202 "sha512" 00:24:13.202 ], 00:24:13.202 "dhchap_dhgroups": [ 00:24:13.202 "null", 00:24:13.202 "ffdhe2048", 00:24:13.202 "ffdhe3072", 00:24:13.202 "ffdhe4096", 00:24:13.202 "ffdhe6144", 00:24:13.202 "ffdhe8192" 00:24:13.202 ] 00:24:13.202 } 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "method": "bdev_nvme_set_hotplug", 00:24:13.202 "params": { 00:24:13.202 "period_us": 100000, 00:24:13.202 "enable": false 00:24:13.202 } 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "method": "bdev_malloc_create", 00:24:13.202 "params": { 00:24:13.202 "name": "malloc0", 00:24:13.202 "num_blocks": 8192, 00:24:13.202 "block_size": 4096, 00:24:13.202 "physical_block_size": 4096, 00:24:13.202 "uuid": "22627714-e34c-4838-9d31-5341266af307", 00:24:13.202 "optimal_io_boundary": 0, 00:24:13.202 "md_size": 0, 00:24:13.202 "dif_type": 0, 00:24:13.202 "dif_is_head_of_md": false, 00:24:13.202 "dif_pi_format": 0 00:24:13.202 } 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "method": "bdev_wait_for_examine" 00:24:13.202 } 00:24:13.202 ] 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "subsystem": "nbd", 00:24:13.202 "config": [] 00:24:13.202 }, 00:24:13.202 { 00:24:13.202 "subsystem": "scheduler", 00:24:13.202 "config": [ 00:24:13.202 { 00:24:13.202 "method": "framework_set_scheduler", 00:24:13.203 "params": { 00:24:13.203 "name": "static" 00:24:13.203 } 00:24:13.203 } 00:24:13.203 ] 00:24:13.203 }, 00:24:13.203 { 00:24:13.203 "subsystem": "nvmf", 00:24:13.203 "config": [ 00:24:13.203 { 00:24:13.203 "method": "nvmf_set_config", 00:24:13.203 "params": { 00:24:13.203 "discovery_filter": "match_any", 00:24:13.203 "admin_cmd_passthru": { 00:24:13.203 "identify_ctrlr": false 00:24:13.203 }, 00:24:13.203 "dhchap_digests": [ 00:24:13.203 "sha256", 00:24:13.203 "sha384", 00:24:13.203 "sha512" 00:24:13.203 ], 00:24:13.203 "dhchap_dhgroups": [ 00:24:13.203 "null", 00:24:13.203 "ffdhe2048", 00:24:13.203 "ffdhe3072", 00:24:13.203 "ffdhe4096", 00:24:13.203 "ffdhe6144", 00:24:13.203 "ffdhe8192" 00:24:13.203 ] 00:24:13.203 } 00:24:13.203 }, 00:24:13.203 { 00:24:13.203 "method": "nvmf_set_max_subsystems", 00:24:13.203 "params": { 00:24:13.203 "max_subsystems": 1024 00:24:13.203 } 00:24:13.203 }, 00:24:13.203 { 00:24:13.203 "method": "nvmf_set_crdt", 00:24:13.203 "params": { 00:24:13.203 "crdt1": 0, 00:24:13.203 "crdt2": 0, 00:24:13.203 "crdt3": 0 00:24:13.203 } 00:24:13.203 }, 00:24:13.203 { 00:24:13.203 "method": "nvmf_create_transport", 00:24:13.203 "params": { 00:24:13.203 "trtype": "TCP", 00:24:13.203 "max_queue_depth": 128, 00:24:13.203 "max_io_qpairs_per_ctrlr": 127, 00:24:13.203 "in_capsule_data_size": 4096, 00:24:13.203 "max_io_size": 131072, 00:24:13.203 "io_unit_size": 131072, 00:24:13.203 "max_aq_depth": 128, 00:24:13.203 "num_shared_buffers": 511, 00:24:13.203 "buf_cache_size": 4294967295, 00:24:13.203 "dif_insert_or_strip": false, 00:24:13.203 "zcopy": false, 00:24:13.203 "c2h_success": false, 00:24:13.203 "sock_priority": 0, 00:24:13.203 "abort_timeout_sec": 1, 00:24:13.203 "ack_timeout": 0, 00:24:13.203 "data_wr_pool_size": 0 00:24:13.203 } 00:24:13.203 }, 00:24:13.203 { 00:24:13.203 "method": "nvmf_create_subsystem", 00:24:13.203 "params": { 00:24:13.203 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:13.203 "allow_any_host": false, 00:24:13.203 "serial_number": "00000000000000000000", 00:24:13.203 "model_number": "SPDK bdev Controller", 00:24:13.203 "max_namespaces": 32, 00:24:13.203 "min_cntlid": 1, 00:24:13.203 "max_cntlid": 65519, 00:24:13.203 "ana_reporting": false 00:24:13.203 } 00:24:13.203 }, 00:24:13.203 { 00:24:13.203 "method": "nvmf_subsystem_add_host", 00:24:13.203 "params": { 00:24:13.203 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:13.203 "host": "nqn.2016-06.io.spdk:host1", 00:24:13.203 "psk": "key0" 00:24:13.203 } 00:24:13.203 }, 00:24:13.203 { 00:24:13.203 "method": "nvmf_subsystem_add_ns", 00:24:13.203 "params": { 00:24:13.203 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:13.203 "namespace": { 00:24:13.203 "nsid": 1, 00:24:13.203 "bdev_name": "malloc0", 00:24:13.203 "nguid": "22627714E34C48389D315341266AF307", 00:24:13.203 "uuid": "22627714-e34c-4838-9d31-5341266af307", 00:24:13.203 "no_auto_visible": false 00:24:13.203 } 00:24:13.203 } 00:24:13.203 }, 00:24:13.203 { 00:24:13.203 "method": "nvmf_subsystem_add_listener", 00:24:13.203 "params": { 00:24:13.203 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:24:13.203 "listen_address": { 00:24:13.203 "trtype": "TCP", 00:24:13.203 "adrfam": "IPv4", 00:24:13.203 "traddr": "10.0.0.2", 00:24:13.203 "trsvcid": "4420" 00:24:13.203 }, 00:24:13.203 "secure_channel": false, 00:24:13.203 "sock_impl": "ssl" 00:24:13.203 } 00:24:13.203 } 00:24:13.203 ] 00:24:13.203 } 00:24:13.203 ] 00:24:13.203 }' 00:24:13.203 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:13.203 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@507 -- # nvmfpid=1921027 00:24:13.203 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@508 -- # waitforlisten 1921027 00:24:13.203 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -c /dev/fd/62 00:24:13.203 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1921027 ']' 00:24:13.203 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:13.203 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:13.203 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:13.203 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:13.203 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:13.203 09:00:03 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:13.203 [2024-11-06 09:00:03.235063] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:24:13.203 [2024-11-06 09:00:03.235119] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:13.203 [2024-11-06 09:00:03.311465] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:13.463 [2024-11-06 09:00:03.344987] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:13.463 [2024-11-06 09:00:03.345024] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:13.463 [2024-11-06 09:00:03.345032] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:13.463 [2024-11-06 09:00:03.345039] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:13.463 [2024-11-06 09:00:03.345044] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:13.463 [2024-11-06 09:00:03.345653] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:13.463 [2024-11-06 09:00:03.545599] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:13.723 [2024-11-06 09:00:03.577618] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:24:13.723 [2024-11-06 09:00:03.577849] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@730 -- # xtrace_disable 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@276 -- # bdevperf_pid=1921376 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@277 -- # waitforlisten 1921376 /var/tmp/bdevperf.sock 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@831 -- # '[' -z 1921376 ']' 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:24:13.984 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@274 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 -c /dev/fd/63 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:13.984 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@274 -- # echo '{ 00:24:13.984 "subsystems": [ 00:24:13.984 { 00:24:13.984 "subsystem": "keyring", 00:24:13.984 "config": [ 00:24:13.984 { 00:24:13.984 "method": "keyring_file_add_key", 00:24:13.984 "params": { 00:24:13.984 "name": "key0", 00:24:13.984 "path": "/tmp/tmp.mY106GdROb" 00:24:13.984 } 00:24:13.984 } 00:24:13.984 ] 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "subsystem": "iobuf", 00:24:13.984 "config": [ 00:24:13.984 { 00:24:13.984 "method": "iobuf_set_options", 00:24:13.984 "params": { 00:24:13.984 "small_pool_count": 8192, 00:24:13.984 "large_pool_count": 1024, 00:24:13.984 "small_bufsize": 8192, 00:24:13.984 "large_bufsize": 135168, 00:24:13.984 "enable_numa": false 00:24:13.984 } 00:24:13.984 } 00:24:13.984 ] 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "subsystem": "sock", 00:24:13.984 "config": [ 00:24:13.984 { 00:24:13.984 "method": "sock_set_default_impl", 00:24:13.984 "params": { 00:24:13.984 "impl_name": "posix" 00:24:13.984 } 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "method": "sock_impl_set_options", 00:24:13.984 "params": { 00:24:13.984 "impl_name": "ssl", 00:24:13.984 "recv_buf_size": 4096, 00:24:13.984 "send_buf_size": 4096, 00:24:13.984 "enable_recv_pipe": true, 00:24:13.984 "enable_quickack": false, 00:24:13.984 "enable_placement_id": 0, 00:24:13.984 "enable_zerocopy_send_server": true, 00:24:13.984 "enable_zerocopy_send_client": false, 00:24:13.984 "zerocopy_threshold": 0, 00:24:13.984 "tls_version": 0, 00:24:13.984 "enable_ktls": false 00:24:13.984 } 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "method": "sock_impl_set_options", 00:24:13.984 "params": { 00:24:13.984 "impl_name": "posix", 00:24:13.984 "recv_buf_size": 2097152, 00:24:13.984 "send_buf_size": 2097152, 00:24:13.984 "enable_recv_pipe": true, 00:24:13.984 "enable_quickack": false, 00:24:13.984 "enable_placement_id": 0, 00:24:13.984 "enable_zerocopy_send_server": true, 00:24:13.984 "enable_zerocopy_send_client": false, 00:24:13.984 "zerocopy_threshold": 0, 00:24:13.984 "tls_version": 0, 00:24:13.984 "enable_ktls": false 00:24:13.984 } 00:24:13.984 } 00:24:13.984 ] 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "subsystem": "vmd", 00:24:13.984 "config": [] 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "subsystem": "accel", 00:24:13.984 "config": [ 00:24:13.984 { 00:24:13.984 "method": "accel_set_options", 00:24:13.984 "params": { 00:24:13.984 "small_cache_size": 128, 00:24:13.984 "large_cache_size": 16, 00:24:13.984 "task_count": 2048, 00:24:13.984 "sequence_count": 2048, 00:24:13.984 "buf_count": 2048 00:24:13.984 } 00:24:13.984 } 00:24:13.984 ] 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "subsystem": "bdev", 00:24:13.984 "config": [ 00:24:13.984 { 00:24:13.984 "method": "bdev_set_options", 00:24:13.984 "params": { 00:24:13.984 "bdev_io_pool_size": 65535, 00:24:13.984 "bdev_io_cache_size": 256, 00:24:13.984 "bdev_auto_examine": true, 00:24:13.984 "iobuf_small_cache_size": 128, 00:24:13.984 "iobuf_large_cache_size": 16 00:24:13.984 } 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "method": "bdev_raid_set_options", 00:24:13.984 "params": { 00:24:13.984 "process_window_size_kb": 1024, 00:24:13.984 "process_max_bandwidth_mb_sec": 0 00:24:13.984 } 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "method": "bdev_iscsi_set_options", 00:24:13.984 "params": { 00:24:13.984 "timeout_sec": 30 00:24:13.984 } 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "method": "bdev_nvme_set_options", 00:24:13.984 "params": { 00:24:13.984 "action_on_timeout": "none", 00:24:13.984 "timeout_us": 0, 00:24:13.984 "timeout_admin_us": 0, 00:24:13.984 "keep_alive_timeout_ms": 10000, 00:24:13.984 "arbitration_burst": 0, 00:24:13.984 "low_priority_weight": 0, 00:24:13.984 "medium_priority_weight": 0, 00:24:13.984 "high_priority_weight": 0, 00:24:13.984 "nvme_adminq_poll_period_us": 10000, 00:24:13.984 "nvme_ioq_poll_period_us": 0, 00:24:13.984 "io_queue_requests": 512, 00:24:13.984 "delay_cmd_submit": true, 00:24:13.984 "transport_retry_count": 4, 00:24:13.984 "bdev_retry_count": 3, 00:24:13.984 "transport_ack_timeout": 0, 00:24:13.984 "ctrlr_loss_timeout_sec": 0, 00:24:13.984 "reconnect_delay_sec": 0, 00:24:13.984 "fast_io_fail_timeout_sec": 0, 00:24:13.984 "disable_auto_failback": false, 00:24:13.984 "generate_uuids": false, 00:24:13.984 "transport_tos": 0, 00:24:13.984 "nvme_error_stat": false, 00:24:13.984 "rdma_srq_size": 0, 00:24:13.984 "io_path_stat": false, 00:24:13.984 "allow_accel_sequence": false, 00:24:13.984 "rdma_max_cq_size": 0, 00:24:13.984 "rdma_cm_event_timeout_ms": 0, 00:24:13.984 "dhchap_digests": [ 00:24:13.984 "sha256", 00:24:13.984 "sha384", 00:24:13.984 "sha512" 00:24:13.984 ], 00:24:13.984 "dhchap_dhgroups": [ 00:24:13.984 "null", 00:24:13.984 "ffdhe2048", 00:24:13.984 "ffdhe3072", 00:24:13.984 "ffdhe4096", 00:24:13.984 "ffdhe6144", 00:24:13.984 "ffdhe8192" 00:24:13.984 ] 00:24:13.984 } 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "method": "bdev_nvme_attach_controller", 00:24:13.984 "params": { 00:24:13.984 "name": "nvme0", 00:24:13.984 "trtype": "TCP", 00:24:13.984 "adrfam": "IPv4", 00:24:13.984 "traddr": "10.0.0.2", 00:24:13.984 "trsvcid": "4420", 00:24:13.984 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:24:13.984 "prchk_reftag": false, 00:24:13.984 "prchk_guard": false, 00:24:13.984 "ctrlr_loss_timeout_sec": 0, 00:24:13.984 "reconnect_delay_sec": 0, 00:24:13.984 "fast_io_fail_timeout_sec": 0, 00:24:13.984 "psk": "key0", 00:24:13.984 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:24:13.984 "hdgst": false, 00:24:13.984 "ddgst": false, 00:24:13.984 "multipath": "multipath" 00:24:13.984 } 00:24:13.984 }, 00:24:13.984 { 00:24:13.984 "method": "bdev_nvme_set_hotplug", 00:24:13.985 "params": { 00:24:13.985 "period_us": 100000, 00:24:13.985 "enable": false 00:24:13.985 } 00:24:13.985 }, 00:24:13.985 { 00:24:13.985 "method": "bdev_enable_histogram", 00:24:13.985 "params": { 00:24:13.985 "name": "nvme0n1", 00:24:13.985 "enable": true 00:24:13.985 } 00:24:13.985 }, 00:24:13.985 { 00:24:13.985 "method": "bdev_wait_for_examine" 00:24:13.985 } 00:24:13.985 ] 00:24:13.985 }, 00:24:13.985 { 00:24:13.985 "subsystem": "nbd", 00:24:13.985 "config": [] 00:24:13.985 } 00:24:13.985 ] 00:24:13.985 }' 00:24:14.245 [2024-11-06 09:00:04.124217] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:24:14.245 [2024-11-06 09:00:04.124271] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1921376 ] 00:24:14.245 [2024-11-06 09:00:04.207203] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:14.245 [2024-11-06 09:00:04.237304] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:14.516 [2024-11-06 09:00:04.373186] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:24:15.092 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:15.092 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@864 -- # return 0 00:24:15.092 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:24:15.092 09:00:04 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # jq -r '.[].name' 00:24:15.092 09:00:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@279 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:24:15.092 09:00:05 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@280 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:24:15.092 Running I/O for 1 seconds... 00:24:16.479 5206.00 IOPS, 20.34 MiB/s 00:24:16.479 Latency(us) 00:24:16.479 [2024-11-06T08:00:06.592Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:16.479 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:24:16.479 Verification LBA range: start 0x0 length 0x2000 00:24:16.479 nvme0n1 : 1.02 5250.34 20.51 0.00 0.00 24217.66 5870.93 34515.63 00:24:16.479 [2024-11-06T08:00:06.592Z] =================================================================================================================== 00:24:16.479 [2024-11-06T08:00:06.592Z] Total : 5250.34 20.51 0.00 0.00 24217.66 5870.93 34515.63 00:24:16.479 { 00:24:16.479 "results": [ 00:24:16.479 { 00:24:16.479 "job": "nvme0n1", 00:24:16.479 "core_mask": "0x2", 00:24:16.479 "workload": "verify", 00:24:16.479 "status": "finished", 00:24:16.479 "verify_range": { 00:24:16.479 "start": 0, 00:24:16.479 "length": 8192 00:24:16.479 }, 00:24:16.479 "queue_depth": 128, 00:24:16.479 "io_size": 4096, 00:24:16.479 "runtime": 1.015935, 00:24:16.479 "iops": 5250.335897473756, 00:24:16.479 "mibps": 20.50912459950686, 00:24:16.479 "io_failed": 0, 00:24:16.479 "io_timeout": 0, 00:24:16.479 "avg_latency_us": 24217.662712160978, 00:24:16.479 "min_latency_us": 5870.933333333333, 00:24:16.479 "max_latency_us": 34515.62666666666 00:24:16.479 } 00:24:16.479 ], 00:24:16.479 "core_count": 1 00:24:16.479 } 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@282 -- # trap - SIGINT SIGTERM EXIT 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@283 -- # cleanup 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@15 -- # process_shm --id 0 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@808 -- # type=--id 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@809 -- # id=0 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@810 -- # '[' --id = --pid ']' 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@814 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@814 -- # shm_files=nvmf_trace.0 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@816 -- # [[ -z nvmf_trace.0 ]] 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@820 -- # for n in $shm_files 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@821 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:24:16.479 nvmf_trace.0 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@823 -- # return 0 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@16 -- # killprocess 1921376 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1921376 ']' 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1921376 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1921376 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1921376' 00:24:16.479 killing process with pid 1921376 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1921376 00:24:16.479 Received shutdown signal, test time was about 1.000000 seconds 00:24:16.479 00:24:16.479 Latency(us) 00:24:16.479 [2024-11-06T08:00:06.592Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:16.479 [2024-11-06T08:00:06.592Z] =================================================================================================================== 00:24:16.479 [2024-11-06T08:00:06.592Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1921376 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@17 -- # nvmftestfini 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@514 -- # nvmfcleanup 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@121 -- # sync 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@124 -- # set +e 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@125 -- # for i in {1..20} 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:24:16.479 rmmod nvme_tcp 00:24:16.479 rmmod nvme_fabrics 00:24:16.479 rmmod nvme_keyring 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@128 -- # set -e 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@129 -- # return 0 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@515 -- # '[' -n 1921027 ']' 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@516 -- # killprocess 1921027 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@950 -- # '[' -z 1921027 ']' 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@954 -- # kill -0 1921027 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # uname 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:16.479 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1921027 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1921027' 00:24:16.740 killing process with pid 1921027 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@969 -- # kill 1921027 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@974 -- # wait 1921027 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@297 -- # iptr 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@789 -- # iptables-save 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@789 -- # iptables-restore 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@302 -- # remove_spdk_ns 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:16.740 09:00:06 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra.nvmf_tls -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra.nvmf_tls -- target/tls.sh@18 -- # rm -f /tmp/tmp.1JPhleGAgO /tmp/tmp.8qvsRXAtTA /tmp/tmp.mY106GdROb 00:24:19.285 00:24:19.285 real 1m21.947s 00:24:19.285 user 2m7.065s 00:24:19.285 sys 0m26.510s 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:24:19.285 ************************************ 00:24:19.285 END TEST nvmf_tls 00:24:19.285 ************************************ 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@42 -- # run_test nvmf_fips /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:24:19.285 ************************************ 00:24:19.285 START TEST nvmf_fips 00:24:19.285 ************************************ 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:24:19.285 * Looking for test storage... 00:24:19.285 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1689 -- # lcov --version 00:24:19.285 09:00:08 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # IFS=.-: 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # read -ra ver1 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # IFS=.-: 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # read -ra ver2 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@338 -- # local 'op=<' 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@340 -- # ver1_l=2 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@341 -- # ver2_l=1 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@344 -- # case "$op" in 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@345 -- # : 1 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 1 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=1 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 1 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=1 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 2 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=2 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 2 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=2 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # return 0 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:19.285 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:24:19.285 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:19.285 --rc genhtml_branch_coverage=1 00:24:19.285 --rc genhtml_function_coverage=1 00:24:19.285 --rc genhtml_legend=1 00:24:19.285 --rc geninfo_all_blocks=1 00:24:19.285 --rc geninfo_unexecuted_blocks=1 00:24:19.285 00:24:19.285 ' 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:24:19.286 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:19.286 --rc genhtml_branch_coverage=1 00:24:19.286 --rc genhtml_function_coverage=1 00:24:19.286 --rc genhtml_legend=1 00:24:19.286 --rc geninfo_all_blocks=1 00:24:19.286 --rc geninfo_unexecuted_blocks=1 00:24:19.286 00:24:19.286 ' 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:24:19.286 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:19.286 --rc genhtml_branch_coverage=1 00:24:19.286 --rc genhtml_function_coverage=1 00:24:19.286 --rc genhtml_legend=1 00:24:19.286 --rc geninfo_all_blocks=1 00:24:19.286 --rc geninfo_unexecuted_blocks=1 00:24:19.286 00:24:19.286 ' 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:24:19.286 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:19.286 --rc genhtml_branch_coverage=1 00:24:19.286 --rc genhtml_function_coverage=1 00:24:19.286 --rc genhtml_legend=1 00:24:19.286 --rc geninfo_all_blocks=1 00:24:19.286 --rc geninfo_unexecuted_blocks=1 00:24:19.286 00:24:19.286 ' 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@7 -- # uname -s 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@15 -- # shopt -s extglob 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@5 -- # export PATH 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@51 -- # : 0 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:19.286 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@90 -- # check_openssl_version 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@84 -- # local target=3.0.0 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # openssl version 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # awk '{print $2}' 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@86 -- # ge 3.1.1 3.0.0 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@376 -- # cmp_versions 3.1.1 '>=' 3.0.0 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # IFS=.-: 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@336 -- # read -ra ver1 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # IFS=.-: 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@337 -- # read -ra ver2 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@338 -- # local 'op=>=' 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@340 -- # ver1_l=3 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@341 -- # ver2_l=3 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@344 -- # case "$op" in 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@348 -- # : 1 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 3 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=3 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 3 =~ ^[0-9]+$ ]] 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 3 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=3 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 3 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=3 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 3 =~ ^[0-9]+$ ]] 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 3 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=3 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v++ )) 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # decimal 1 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=1 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 1 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@365 -- # ver1[v]=1 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # decimal 0 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@353 -- # local d=0 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@354 -- # [[ 0 =~ ^[0-9]+$ ]] 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@355 -- # echo 0 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@366 -- # ver2[v]=0 00:24:19.286 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- scripts/common.sh@367 -- # return 0 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@96 -- # openssl info -modulesdir 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@96 -- # [[ ! -f /usr/lib64/ossl-modules/fips.so ]] 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@101 -- # openssl fipsinstall -help 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@101 -- # warn='This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode' 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@102 -- # [[ This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode == \T\h\i\s\ \c\o\m\m\a\n\d\ \i\s\ \n\o\t\ \e\n\a\b\l\e\d* ]] 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@105 -- # export callback=build_openssl_config 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@105 -- # callback=build_openssl_config 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@114 -- # build_openssl_config 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@38 -- # cat 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@58 -- # [[ ! -t 0 ]] 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@59 -- # cat - 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@115 -- # export OPENSSL_CONF=spdk_fips.conf 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@115 -- # OPENSSL_CONF=spdk_fips.conf 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # mapfile -t providers 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # openssl list -providers 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@117 -- # grep name 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # (( 2 != 2 )) 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # [[ name: openssl base provider != *base* ]] 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@121 -- # [[ name: red hat enterprise linux 9 - openssl fips provider != *fips* ]] 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@128 -- # NOT openssl md5 /dev/fd/62 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@650 -- # local es=0 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@652 -- # valid_exec_arg openssl md5 /dev/fd/62 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@128 -- # : 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@638 -- # local arg=openssl 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@642 -- # type -t openssl 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # type -P openssl 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # arg=/usr/bin/openssl 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@644 -- # [[ -x /usr/bin/openssl ]] 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@653 -- # openssl md5 /dev/fd/62 00:24:19.287 Error setting digest 00:24:19.287 4012BCAA687F0000:error:0308010C:digital envelope routines:inner_evp_generic_fetch:unsupported:crypto/evp/evp_fetch.c:341:Global default library context, Algorithm (MD5 : 95), Properties () 00:24:19.287 4012BCAA687F0000:error:03000086:digital envelope routines:evp_md_init_internal:initialization error:crypto/evp/digest.c:272: 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@653 -- # es=1 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@131 -- # nvmftestinit 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@474 -- # prepare_net_devs 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@436 -- # local -g is_hw=no 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@438 -- # remove_spdk_ns 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@309 -- # xtrace_disable 00:24:19.287 09:00:09 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@315 -- # pci_devs=() 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@315 -- # local -a pci_devs 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@316 -- # pci_net_devs=() 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@317 -- # pci_drivers=() 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@317 -- # local -A pci_drivers 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@319 -- # net_devs=() 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@319 -- # local -ga net_devs 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@320 -- # e810=() 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@320 -- # local -ga e810 00:24:27.435 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@321 -- # x722=() 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@321 -- # local -ga x722 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@322 -- # mlx=() 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@322 -- # local -ga mlx 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:24:27.436 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:24:27.436 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@416 -- # [[ up == up ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:24:27.436 Found net devices under 0000:4b:00.0: cvl_0_0 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@416 -- # [[ up == up ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:24:27.436 Found net devices under 0000:4b:00.1: cvl_0_1 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@440 -- # is_hw=yes 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:24:27.436 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:27.436 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.661 ms 00:24:27.436 00:24:27.436 --- 10.0.0.2 ping statistics --- 00:24:27.436 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:27.436 rtt min/avg/max/mdev = 0.661/0.661/0.661/0.000 ms 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:27.436 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:27.436 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.214 ms 00:24:27.436 00:24:27.436 --- 10.0.0.1 ping statistics --- 00:24:27.436 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:27.436 rtt min/avg/max/mdev = 0.214/0.214/0.214/0.000 ms 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@448 -- # return 0 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@132 -- # nvmfappstart -m 0x2 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@724 -- # xtrace_disable 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@507 -- # nvmfpid=1926537 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@508 -- # waitforlisten 1926537 00:24:27.436 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:24:27.437 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@831 -- # '[' -z 1926537 ']' 00:24:27.437 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:27.437 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:27.437 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:27.437 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:27.437 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:27.437 09:00:16 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:24:27.437 [2024-11-06 09:00:16.579836] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:24:27.437 [2024-11-06 09:00:16.579907] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:27.437 [2024-11-06 09:00:16.679876] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:27.437 [2024-11-06 09:00:16.728858] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:27.437 [2024-11-06 09:00:16.728908] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:27.437 [2024-11-06 09:00:16.728917] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:27.437 [2024-11-06 09:00:16.728925] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:27.437 [2024-11-06 09:00:16.728931] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:27.437 [2024-11-06 09:00:16.729713] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@864 -- # return 0 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@730 -- # xtrace_disable 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@134 -- # trap cleanup EXIT 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@137 -- # key=NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@138 -- # mktemp -t spdk-psk.XXX 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@138 -- # key_path=/tmp/spdk-psk.sTf 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@139 -- # echo -n NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@140 -- # chmod 0600 /tmp/spdk-psk.sTf 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@142 -- # setup_nvmf_tgt_conf /tmp/spdk-psk.sTf 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@22 -- # local key=/tmp/spdk-psk.sTf 00:24:27.437 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:24:27.699 [2024-11-06 09:00:17.609206] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:27.699 [2024-11-06 09:00:17.625215] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:24:27.699 [2024-11-06 09:00:17.625521] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:27.699 malloc0 00:24:27.699 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@145 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:24:27.699 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@148 -- # bdevperf_pid=1926723 00:24:27.699 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@149 -- # waitforlisten 1926723 /var/tmp/bdevperf.sock 00:24:27.699 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@146 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:24:27.699 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@831 -- # '[' -z 1926723 ']' 00:24:27.699 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:24:27.699 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:27.699 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:24:27.699 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:24:27.699 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:27.699 09:00:17 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:24:27.699 [2024-11-06 09:00:17.769477] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:24:27.699 [2024-11-06 09:00:17.769556] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1926723 ] 00:24:27.960 [2024-11-06 09:00:17.832622] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:27.960 [2024-11-06 09:00:17.869737] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:24:28.532 09:00:18 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:28.532 09:00:18 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@864 -- # return 0 00:24:28.532 09:00:18 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@151 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/spdk-psk.sTf 00:24:28.792 09:00:18 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@152 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk key0 00:24:28.792 [2024-11-06 09:00:18.886004] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:24:29.053 TLSTESTn1 00:24:29.053 09:00:18 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@156 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:24:29.053 Running I/O for 10 seconds... 00:24:31.381 5401.00 IOPS, 21.10 MiB/s [2024-11-06T08:00:22.433Z] 5142.00 IOPS, 20.09 MiB/s [2024-11-06T08:00:23.373Z] 4911.00 IOPS, 19.18 MiB/s [2024-11-06T08:00:24.391Z] 5220.75 IOPS, 20.39 MiB/s [2024-11-06T08:00:25.331Z] 5194.20 IOPS, 20.29 MiB/s [2024-11-06T08:00:26.405Z] 5153.67 IOPS, 20.13 MiB/s [2024-11-06T08:00:27.347Z] 5189.57 IOPS, 20.27 MiB/s [2024-11-06T08:00:28.290Z] 5360.25 IOPS, 20.94 MiB/s [2024-11-06T08:00:29.232Z] 5422.56 IOPS, 21.18 MiB/s [2024-11-06T08:00:29.232Z] 5429.00 IOPS, 21.21 MiB/s 00:24:39.119 Latency(us) 00:24:39.119 [2024-11-06T08:00:29.232Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:39.119 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:24:39.119 Verification LBA range: start 0x0 length 0x2000 00:24:39.119 TLSTESTn1 : 10.03 5427.97 21.20 0.00 0.00 23541.18 6034.77 53739.52 00:24:39.119 [2024-11-06T08:00:29.232Z] =================================================================================================================== 00:24:39.119 [2024-11-06T08:00:29.232Z] Total : 5427.97 21.20 0.00 0.00 23541.18 6034.77 53739.52 00:24:39.119 { 00:24:39.119 "results": [ 00:24:39.119 { 00:24:39.119 "job": "TLSTESTn1", 00:24:39.119 "core_mask": "0x4", 00:24:39.119 "workload": "verify", 00:24:39.119 "status": "finished", 00:24:39.119 "verify_range": { 00:24:39.119 "start": 0, 00:24:39.119 "length": 8192 00:24:39.119 }, 00:24:39.119 "queue_depth": 128, 00:24:39.119 "io_size": 4096, 00:24:39.119 "runtime": 10.025472, 00:24:39.119 "iops": 5427.973864971145, 00:24:39.119 "mibps": 21.203022910043536, 00:24:39.119 "io_failed": 0, 00:24:39.119 "io_timeout": 0, 00:24:39.119 "avg_latency_us": 23541.184315973882, 00:24:39.119 "min_latency_us": 6034.7733333333335, 00:24:39.119 "max_latency_us": 53739.52 00:24:39.119 } 00:24:39.119 ], 00:24:39.119 "core_count": 1 00:24:39.119 } 00:24:39.119 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@1 -- # cleanup 00:24:39.119 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@15 -- # process_shm --id 0 00:24:39.119 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@808 -- # type=--id 00:24:39.119 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@809 -- # id=0 00:24:39.119 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@810 -- # '[' --id = --pid ']' 00:24:39.119 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@814 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:24:39.119 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@814 -- # shm_files=nvmf_trace.0 00:24:39.119 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@816 -- # [[ -z nvmf_trace.0 ]] 00:24:39.119 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@820 -- # for n in $shm_files 00:24:39.119 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@821 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:24:39.119 nvmf_trace.0 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@823 -- # return 0 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@16 -- # killprocess 1926723 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@950 -- # '[' -z 1926723 ']' 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@954 -- # kill -0 1926723 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@955 -- # uname 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1926723 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1926723' 00:24:39.380 killing process with pid 1926723 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@969 -- # kill 1926723 00:24:39.380 Received shutdown signal, test time was about 10.000000 seconds 00:24:39.380 00:24:39.380 Latency(us) 00:24:39.380 [2024-11-06T08:00:29.493Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:39.380 [2024-11-06T08:00:29.493Z] =================================================================================================================== 00:24:39.380 [2024-11-06T08:00:29.493Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@974 -- # wait 1926723 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@17 -- # nvmftestfini 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@514 -- # nvmfcleanup 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@121 -- # sync 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:24:39.380 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@124 -- # set +e 00:24:39.381 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@125 -- # for i in {1..20} 00:24:39.381 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:24:39.381 rmmod nvme_tcp 00:24:39.381 rmmod nvme_fabrics 00:24:39.381 rmmod nvme_keyring 00:24:39.381 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:24:39.381 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@128 -- # set -e 00:24:39.381 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@129 -- # return 0 00:24:39.381 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@515 -- # '[' -n 1926537 ']' 00:24:39.381 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@516 -- # killprocess 1926537 00:24:39.381 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@950 -- # '[' -z 1926537 ']' 00:24:39.381 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@954 -- # kill -0 1926537 00:24:39.381 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@955 -- # uname 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1926537 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1926537' 00:24:39.642 killing process with pid 1926537 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@969 -- # kill 1926537 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@974 -- # wait 1926537 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@297 -- # iptr 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@789 -- # iptables-save 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@789 -- # iptables-restore 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@302 -- # remove_spdk_ns 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:39.642 09:00:29 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:42.190 09:00:31 nvmf_tcp.nvmf_target_extra.nvmf_fips -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:24:42.190 09:00:31 nvmf_tcp.nvmf_target_extra.nvmf_fips -- fips/fips.sh@18 -- # rm -f /tmp/spdk-psk.sTf 00:24:42.190 00:24:42.190 real 0m22.857s 00:24:42.190 user 0m24.607s 00:24:42.190 sys 0m9.526s 00:24:42.191 09:00:31 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:42.191 09:00:31 nvmf_tcp.nvmf_target_extra.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:24:42.191 ************************************ 00:24:42.191 END TEST nvmf_fips 00:24:42.191 ************************************ 00:24:42.191 09:00:31 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@43 -- # run_test nvmf_control_msg_list /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/control_msg_list.sh --transport=tcp 00:24:42.191 09:00:31 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:24:42.191 09:00:31 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:42.191 09:00:31 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:24:42.191 ************************************ 00:24:42.191 START TEST nvmf_control_msg_list 00:24:42.191 ************************************ 00:24:42.191 09:00:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/control_msg_list.sh --transport=tcp 00:24:42.191 * Looking for test storage... 00:24:42.191 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:24:42.191 09:00:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:24:42.191 09:00:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1689 -- # lcov --version 00:24:42.191 09:00:31 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@336 -- # IFS=.-: 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@336 -- # read -ra ver1 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@337 -- # IFS=.-: 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@337 -- # read -ra ver2 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@338 -- # local 'op=<' 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@340 -- # ver1_l=2 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@341 -- # ver2_l=1 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@344 -- # case "$op" in 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@345 -- # : 1 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@365 -- # decimal 1 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@353 -- # local d=1 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@355 -- # echo 1 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@365 -- # ver1[v]=1 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@366 -- # decimal 2 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@353 -- # local d=2 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@355 -- # echo 2 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@366 -- # ver2[v]=2 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@368 -- # return 0 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:24:42.191 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:42.191 --rc genhtml_branch_coverage=1 00:24:42.191 --rc genhtml_function_coverage=1 00:24:42.191 --rc genhtml_legend=1 00:24:42.191 --rc geninfo_all_blocks=1 00:24:42.191 --rc geninfo_unexecuted_blocks=1 00:24:42.191 00:24:42.191 ' 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:24:42.191 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:42.191 --rc genhtml_branch_coverage=1 00:24:42.191 --rc genhtml_function_coverage=1 00:24:42.191 --rc genhtml_legend=1 00:24:42.191 --rc geninfo_all_blocks=1 00:24:42.191 --rc geninfo_unexecuted_blocks=1 00:24:42.191 00:24:42.191 ' 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:24:42.191 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:42.191 --rc genhtml_branch_coverage=1 00:24:42.191 --rc genhtml_function_coverage=1 00:24:42.191 --rc genhtml_legend=1 00:24:42.191 --rc geninfo_all_blocks=1 00:24:42.191 --rc geninfo_unexecuted_blocks=1 00:24:42.191 00:24:42.191 ' 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:24:42.191 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:42.191 --rc genhtml_branch_coverage=1 00:24:42.191 --rc genhtml_function_coverage=1 00:24:42.191 --rc genhtml_legend=1 00:24:42.191 --rc geninfo_all_blocks=1 00:24:42.191 --rc geninfo_unexecuted_blocks=1 00:24:42.191 00:24:42.191 ' 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@7 -- # uname -s 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@15 -- # shopt -s extglob 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:42.191 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@5 -- # export PATH 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@51 -- # : 0 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:42.192 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@12 -- # nvmftestinit 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@474 -- # prepare_net_devs 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@436 -- # local -g is_hw=no 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@438 -- # remove_spdk_ns 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@309 -- # xtrace_disable 00:24:42.192 09:00:32 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@315 -- # pci_devs=() 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@315 -- # local -a pci_devs 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@316 -- # pci_net_devs=() 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@317 -- # pci_drivers=() 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@317 -- # local -A pci_drivers 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@319 -- # net_devs=() 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@319 -- # local -ga net_devs 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@320 -- # e810=() 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@320 -- # local -ga e810 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@321 -- # x722=() 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@321 -- # local -ga x722 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@322 -- # mlx=() 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@322 -- # local -ga mlx 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:24:50.334 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:24:50.334 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@416 -- # [[ up == up ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:24:50.334 Found net devices under 0000:4b:00.0: cvl_0_0 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@416 -- # [[ up == up ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:24:50.334 Found net devices under 0000:4b:00.1: cvl_0_1 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@440 -- # is_hw=yes 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:50.334 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:50.335 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:24:50.335 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:24:50.335 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:24:50.335 09:00:38 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:24:50.335 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:50.335 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.619 ms 00:24:50.335 00:24:50.335 --- 10.0.0.2 ping statistics --- 00:24:50.335 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:50.335 rtt min/avg/max/mdev = 0.619/0.619/0.619/0.000 ms 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:50.335 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:50.335 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.265 ms 00:24:50.335 00:24:50.335 --- 10.0.0.1 ping statistics --- 00:24:50.335 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:50.335 rtt min/avg/max/mdev = 0.265/0.265/0.265/0.000 ms 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@448 -- # return 0 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@13 -- # nvmfappstart 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@724 -- # xtrace_disable 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@507 -- # nvmfpid=1933241 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@508 -- # waitforlisten 1933241 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@831 -- # '[' -z 1933241 ']' 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@836 -- # local max_retries=100 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:50.335 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@840 -- # xtrace_disable 00:24:50.335 09:00:39 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:24:50.335 [2024-11-06 09:00:39.363291] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:24:50.335 [2024-11-06 09:00:39.363362] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:50.335 [2024-11-06 09:00:39.445305] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:50.335 [2024-11-06 09:00:39.485470] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:50.335 [2024-11-06 09:00:39.485508] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:50.335 [2024-11-06 09:00:39.485516] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:50.335 [2024-11-06 09:00:39.485523] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:50.335 [2024-11-06 09:00:39.485529] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:50.335 [2024-11-06 09:00:39.486129] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@864 -- # return 0 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@730 -- # xtrace_disable 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@15 -- # subnqn=nqn.2024-07.io.spdk:cnode0 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@16 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@19 -- # rpc_cmd nvmf_create_transport '-t tcp -o' --in-capsule-data-size 768 --control-msg-num 1 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:24:50.335 [2024-11-06 09:00:40.189013] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2024-07.io.spdk:cnode0 -a 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@21 -- # rpc_cmd bdev_malloc_create -b Malloc0 32 512 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:24:50.335 Malloc0 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2024-07.io.spdk:cnode0 Malloc0 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2024-07.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@561 -- # xtrace_disable 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:24:50.335 [2024-11-06 09:00:40.239821] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@27 -- # perf_pid1=1933280 00:24:50.335 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x2 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:24:50.336 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@29 -- # perf_pid2=1933281 00:24:50.336 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x4 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:24:50.336 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@31 -- # perf_pid3=1933283 00:24:50.336 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@33 -- # wait 1933280 00:24:50.336 09:00:40 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x8 -q 1 -o 4096 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:24:50.336 [2024-11-06 09:00:40.310242] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:24:50.336 [2024-11-06 09:00:40.340333] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:24:50.336 [2024-11-06 09:00:40.340615] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:24:51.719 Initializing NVMe Controllers 00:24:51.719 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:24:51.719 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 2 00:24:51.719 Initialization complete. Launching workers. 00:24:51.719 ======================================================== 00:24:51.719 Latency(us) 00:24:51.719 Device Information : IOPS MiB/s Average min max 00:24:51.719 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 2: 1745.00 6.82 573.17 159.19 858.16 00:24:51.719 ======================================================== 00:24:51.719 Total : 1745.00 6.82 573.17 159.19 858.16 00:24:51.719 00:24:51.719 Initializing NVMe Controllers 00:24:51.719 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:24:51.719 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 1 00:24:51.719 Initialization complete. Launching workers. 00:24:51.719 ======================================================== 00:24:51.719 Latency(us) 00:24:51.719 Device Information : IOPS MiB/s Average min max 00:24:51.719 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 1: 25.00 0.10 40909.83 40669.54 41077.34 00:24:51.719 ======================================================== 00:24:51.719 Total : 25.00 0.10 40909.83 40669.54 41077.34 00:24:51.719 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@34 -- # wait 1933281 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@35 -- # wait 1933283 00:24:51.719 Initializing NVMe Controllers 00:24:51.719 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:24:51.719 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 3 00:24:51.719 Initialization complete. Launching workers. 00:24:51.719 ======================================================== 00:24:51.719 Latency(us) 00:24:51.719 Device Information : IOPS MiB/s Average min max 00:24:51.719 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 3: 25.00 0.10 40900.49 40787.93 40966.58 00:24:51.719 ======================================================== 00:24:51.719 Total : 25.00 0.10 40900.49 40787.93 40966.58 00:24:51.719 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- target/control_msg_list.sh@38 -- # nvmftestfini 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@514 -- # nvmfcleanup 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@121 -- # sync 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@124 -- # set +e 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@125 -- # for i in {1..20} 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:24:51.719 rmmod nvme_tcp 00:24:51.719 rmmod nvme_fabrics 00:24:51.719 rmmod nvme_keyring 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@128 -- # set -e 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@129 -- # return 0 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@515 -- # '[' -n 1933241 ']' 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@516 -- # killprocess 1933241 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@950 -- # '[' -z 1933241 ']' 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@954 -- # kill -0 1933241 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@955 -- # uname 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1933241 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1933241' 00:24:51.719 killing process with pid 1933241 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@969 -- # kill 1933241 00:24:51.719 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@974 -- # wait 1933241 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@297 -- # iptr 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@789 -- # iptables-save 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@789 -- # iptables-restore 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@302 -- # remove_spdk_ns 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:51.980 09:00:41 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:53.892 09:00:43 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:24:53.892 00:24:53.892 real 0m12.110s 00:24:53.892 user 0m8.130s 00:24:53.892 sys 0m6.183s 00:24:53.892 09:00:43 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@1126 -- # xtrace_disable 00:24:53.893 09:00:43 nvmf_tcp.nvmf_target_extra.nvmf_control_msg_list -- common/autotest_common.sh@10 -- # set +x 00:24:53.893 ************************************ 00:24:53.893 END TEST nvmf_control_msg_list 00:24:53.893 ************************************ 00:24:53.893 09:00:43 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@44 -- # run_test nvmf_wait_for_buf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/wait_for_buf.sh --transport=tcp 00:24:53.893 09:00:43 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:24:53.893 09:00:43 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:24:53.893 09:00:43 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:24:54.155 ************************************ 00:24:54.155 START TEST nvmf_wait_for_buf 00:24:54.155 ************************************ 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/wait_for_buf.sh --transport=tcp 00:24:54.155 * Looking for test storage... 00:24:54.155 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1689 -- # lcov --version 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@336 -- # IFS=.-: 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@336 -- # read -ra ver1 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@337 -- # IFS=.-: 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@337 -- # read -ra ver2 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@338 -- # local 'op=<' 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@340 -- # ver1_l=2 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@341 -- # ver2_l=1 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@344 -- # case "$op" in 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@345 -- # : 1 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@364 -- # (( v = 0 )) 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@365 -- # decimal 1 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@353 -- # local d=1 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@355 -- # echo 1 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@365 -- # ver1[v]=1 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@366 -- # decimal 2 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@353 -- # local d=2 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@355 -- # echo 2 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@366 -- # ver2[v]=2 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@368 -- # return 0 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:24:54.155 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:54.155 --rc genhtml_branch_coverage=1 00:24:54.155 --rc genhtml_function_coverage=1 00:24:54.155 --rc genhtml_legend=1 00:24:54.155 --rc geninfo_all_blocks=1 00:24:54.155 --rc geninfo_unexecuted_blocks=1 00:24:54.155 00:24:54.155 ' 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:24:54.155 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:54.155 --rc genhtml_branch_coverage=1 00:24:54.155 --rc genhtml_function_coverage=1 00:24:54.155 --rc genhtml_legend=1 00:24:54.155 --rc geninfo_all_blocks=1 00:24:54.155 --rc geninfo_unexecuted_blocks=1 00:24:54.155 00:24:54.155 ' 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:24:54.155 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:54.155 --rc genhtml_branch_coverage=1 00:24:54.155 --rc genhtml_function_coverage=1 00:24:54.155 --rc genhtml_legend=1 00:24:54.155 --rc geninfo_all_blocks=1 00:24:54.155 --rc geninfo_unexecuted_blocks=1 00:24:54.155 00:24:54.155 ' 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:24:54.155 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:24:54.155 --rc genhtml_branch_coverage=1 00:24:54.155 --rc genhtml_function_coverage=1 00:24:54.155 --rc genhtml_legend=1 00:24:54.155 --rc geninfo_all_blocks=1 00:24:54.155 --rc geninfo_unexecuted_blocks=1 00:24:54.155 00:24:54.155 ' 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@7 -- # uname -s 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@15 -- # shopt -s extglob 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:54.155 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@5 -- # export PATH 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@51 -- # : 0 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:24:54.156 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:24:54.156 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@12 -- # nvmftestinit 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@474 -- # prepare_net_devs 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@436 -- # local -g is_hw=no 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@438 -- # remove_spdk_ns 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@309 -- # xtrace_disable 00:24:54.418 09:00:44 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@315 -- # pci_devs=() 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@319 -- # net_devs=() 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@320 -- # e810=() 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@320 -- # local -ga e810 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@321 -- # x722=() 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@321 -- # local -ga x722 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@322 -- # mlx=() 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@322 -- # local -ga mlx 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:25:02.561 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:25:02.561 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:02.561 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:25:02.561 Found net devices under 0000:4b:00.0: cvl_0_0 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:25:02.562 Found net devices under 0000:4b:00.1: cvl_0_1 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@440 -- # is_hw=yes 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:25:02.562 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:02.562 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.639 ms 00:25:02.562 00:25:02.562 --- 10.0.0.2 ping statistics --- 00:25:02.562 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:02.562 rtt min/avg/max/mdev = 0.639/0.639/0.639/0.000 ms 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:02.562 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:02.562 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.324 ms 00:25:02.562 00:25:02.562 --- 10.0.0.1 ping statistics --- 00:25:02.562 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:02.562 rtt min/avg/max/mdev = 0.324/0.324/0.324/0.000 ms 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@448 -- # return 0 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@13 -- # nvmfappstart --wait-for-rpc 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@724 -- # xtrace_disable 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@507 -- # nvmfpid=1937922 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@508 -- # waitforlisten 1937922 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@831 -- # '[' -z 1937922 ']' 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:02.562 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:02.562 09:00:51 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.562 [2024-11-06 09:00:51.815240] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:25:02.562 [2024-11-06 09:00:51.815310] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:02.562 [2024-11-06 09:00:51.896970] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:02.562 [2024-11-06 09:00:51.937079] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:02.562 [2024-11-06 09:00:51.937117] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:02.562 [2024-11-06 09:00:51.937124] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:02.562 [2024-11-06 09:00:51.937131] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:02.562 [2024-11-06 09:00:51.937137] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:02.562 [2024-11-06 09:00:51.937719] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:02.562 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:02.562 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@864 -- # return 0 00:25:02.562 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:25:02.562 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@730 -- # xtrace_disable 00:25:02.562 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.562 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:02.562 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@15 -- # subnqn=nqn.2024-07.io.spdk:cnode0 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@16 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@19 -- # rpc_cmd accel_set_options --small-cache-size 0 --large-cache-size 0 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@20 -- # rpc_cmd iobuf_set_options --small-pool-count 154 --small_bufsize=8192 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@21 -- # rpc_cmd framework_start_init 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:02.563 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@22 -- # rpc_cmd bdev_malloc_create -b Malloc0 32 512 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.824 Malloc0 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@23 -- # rpc_cmd nvmf_create_transport '-t tcp -o' -u 8192 -n 24 -b 24 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.824 [2024-11-06 09:00:52.737518] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2024-07.io.spdk:cnode0 -a -s SPDK00000000000001 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2024-07.io.spdk:cnode0 Malloc0 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2024-07.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:02.824 [2024-11-06 09:00:52.773738] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:02.824 09:00:52 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 4 -o 131072 -w randread -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:25:02.824 [2024-11-06 09:00:52.875830] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:25:04.208 Initializing NVMe Controllers 00:25:04.208 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2024-07.io.spdk:cnode0 00:25:04.208 Associating TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 with lcore 0 00:25:04.208 Initialization complete. Launching workers. 00:25:04.208 ======================================================== 00:25:04.208 Latency(us) 00:25:04.208 Device Information : IOPS MiB/s Average min max 00:25:04.208 TCP (addr:10.0.0.2 subnqn:nqn.2024-07.io.spdk:cnode0) NSID 1 from core 0: 129.00 16.12 32295.06 7990.65 63851.46 00:25:04.208 ======================================================== 00:25:04.208 Total : 129.00 16.12 32295.06 7990.65 63851.46 00:25:04.208 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # rpc_cmd iobuf_get_stats 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # jq -r '.[] | select(.module == "nvmf_TCP") | .small_pool.retry' 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@32 -- # retry_count=2038 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@33 -- # [[ 2038 -eq 0 ]] 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- target/wait_for_buf.sh@38 -- # nvmftestfini 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@514 -- # nvmfcleanup 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@121 -- # sync 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@124 -- # set +e 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@125 -- # for i in {1..20} 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:25:04.468 rmmod nvme_tcp 00:25:04.468 rmmod nvme_fabrics 00:25:04.468 rmmod nvme_keyring 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@128 -- # set -e 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@129 -- # return 0 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@515 -- # '[' -n 1937922 ']' 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@516 -- # killprocess 1937922 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@950 -- # '[' -z 1937922 ']' 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@954 -- # kill -0 1937922 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@955 -- # uname 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1937922 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1937922' 00:25:04.468 killing process with pid 1937922 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@969 -- # kill 1937922 00:25:04.468 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@974 -- # wait 1937922 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@297 -- # iptr 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@789 -- # iptables-save 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@789 -- # iptables-restore 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:04.729 09:00:54 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:06.643 09:00:56 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:25:06.643 00:25:06.643 real 0m12.696s 00:25:06.643 user 0m5.208s 00:25:06.643 sys 0m6.046s 00:25:06.643 09:00:56 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:25:06.643 09:00:56 nvmf_tcp.nvmf_target_extra.nvmf_wait_for_buf -- common/autotest_common.sh@10 -- # set +x 00:25:06.643 ************************************ 00:25:06.643 END TEST nvmf_wait_for_buf 00:25:06.643 ************************************ 00:25:06.904 09:00:56 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@47 -- # '[' 0 -eq 1 ']' 00:25:06.904 09:00:56 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@53 -- # [[ phy == phy ]] 00:25:06.904 09:00:56 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@54 -- # '[' tcp = tcp ']' 00:25:06.904 09:00:56 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@55 -- # gather_supported_nvmf_pci_devs 00:25:06.904 09:00:56 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@309 -- # xtrace_disable 00:25:06.904 09:00:56 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@315 -- # pci_devs=() 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@319 -- # net_devs=() 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@320 -- # e810=() 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@320 -- # local -ga e810 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@321 -- # x722=() 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@321 -- # local -ga x722 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@322 -- # mlx=() 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@322 -- # local -ga mlx 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:25:15.048 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:25:15.048 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:25:15.048 Found net devices under 0000:4b:00.0: cvl_0_0 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:25:15.048 Found net devices under 0000:4b:00.1: cvl_0_1 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@56 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@57 -- # (( 2 > 0 )) 00:25:15.048 09:01:03 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@58 -- # run_test nvmf_perf_adq /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:25:15.049 ************************************ 00:25:15.049 START TEST nvmf_perf_adq 00:25:15.049 ************************************ 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:25:15.049 * Looking for test storage... 00:25:15.049 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1689 -- # lcov --version 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@333 -- # local ver1 ver1_l 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@334 -- # local ver2 ver2_l 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@336 -- # IFS=.-: 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@336 -- # read -ra ver1 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@337 -- # IFS=.-: 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@337 -- # read -ra ver2 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@338 -- # local 'op=<' 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@340 -- # ver1_l=2 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@341 -- # ver2_l=1 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@344 -- # case "$op" in 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@345 -- # : 1 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@364 -- # (( v = 0 )) 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@365 -- # decimal 1 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@353 -- # local d=1 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@355 -- # echo 1 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@365 -- # ver1[v]=1 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@366 -- # decimal 2 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@353 -- # local d=2 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@355 -- # echo 2 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@366 -- # ver2[v]=2 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@368 -- # return 0 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:25:15.049 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:15.049 --rc genhtml_branch_coverage=1 00:25:15.049 --rc genhtml_function_coverage=1 00:25:15.049 --rc genhtml_legend=1 00:25:15.049 --rc geninfo_all_blocks=1 00:25:15.049 --rc geninfo_unexecuted_blocks=1 00:25:15.049 00:25:15.049 ' 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:25:15.049 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:15.049 --rc genhtml_branch_coverage=1 00:25:15.049 --rc genhtml_function_coverage=1 00:25:15.049 --rc genhtml_legend=1 00:25:15.049 --rc geninfo_all_blocks=1 00:25:15.049 --rc geninfo_unexecuted_blocks=1 00:25:15.049 00:25:15.049 ' 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:25:15.049 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:15.049 --rc genhtml_branch_coverage=1 00:25:15.049 --rc genhtml_function_coverage=1 00:25:15.049 --rc genhtml_legend=1 00:25:15.049 --rc geninfo_all_blocks=1 00:25:15.049 --rc geninfo_unexecuted_blocks=1 00:25:15.049 00:25:15.049 ' 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:25:15.049 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:25:15.049 --rc genhtml_branch_coverage=1 00:25:15.049 --rc genhtml_function_coverage=1 00:25:15.049 --rc genhtml_legend=1 00:25:15.049 --rc geninfo_all_blocks=1 00:25:15.049 --rc geninfo_unexecuted_blocks=1 00:25:15.049 00:25:15.049 ' 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@7 -- # uname -s 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@15 -- # shopt -s extglob 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@5 -- # export PATH 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@51 -- # : 0 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:25:15.049 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:25:15.049 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:25:15.050 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:25:15.050 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@55 -- # have_pci_nics=0 00:25:15.050 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@11 -- # gather_supported_nvmf_pci_devs 00:25:15.050 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:25:15.050 09:01:03 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:21.638 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:21.638 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:25:21.639 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:25:21.639 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:25:21.639 Found net devices under 0000:4b:00.0: cvl_0_0 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:25:21.639 Found net devices under 0000:4b:00.1: cvl_0_1 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@12 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@13 -- # (( 2 == 0 )) 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@18 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@68 -- # adq_reload_driver 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@58 -- # modprobe -a sch_mqprio 00:25:21.639 09:01:10 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@61 -- # rmmod ice 00:25:22.212 09:01:12 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@62 -- # modprobe ice 00:25:24.138 09:01:14 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@63 -- # sleep 5 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@76 -- # nvmftestinit 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@474 -- # prepare_net_devs 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@436 -- # local -g is_hw=no 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@438 -- # remove_spdk_ns 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:25:29.422 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:25:29.422 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:25:29.422 Found net devices under 0000:4b:00.0: cvl_0_0 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:25:29.422 Found net devices under 0000:4b:00.1: cvl_0_1 00:25:29.422 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # is_hw=yes 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:25:29.423 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:29.423 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.616 ms 00:25:29.423 00:25:29.423 --- 10.0.0.2 ping statistics --- 00:25:29.423 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:29.423 rtt min/avg/max/mdev = 0.616/0.616/0.616/0.000 ms 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:29.423 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:29.423 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.285 ms 00:25:29.423 00:25:29.423 --- 10.0.0.1 ping statistics --- 00:25:29.423 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:29.423 rtt min/avg/max/mdev = 0.285/0.285/0.285/0.000 ms 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@448 -- # return 0 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@77 -- # nvmfappstart -m 0xF --wait-for-rpc 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@724 -- # xtrace_disable 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@507 -- # nvmfpid=1948046 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@508 -- # waitforlisten 1948046 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@831 -- # '[' -z 1948046 ']' 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:29.423 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:29.423 09:01:19 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:29.687 [2024-11-06 09:01:19.563155] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:25:29.687 [2024-11-06 09:01:19.563209] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:29.687 [2024-11-06 09:01:19.644524] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:25:29.688 [2024-11-06 09:01:19.685034] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:29.688 [2024-11-06 09:01:19.685070] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:29.688 [2024-11-06 09:01:19.685078] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:29.688 [2024-11-06 09:01:19.685086] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:29.688 [2024-11-06 09:01:19.685091] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:29.688 [2024-11-06 09:01:19.686647] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:29.688 [2024-11-06 09:01:19.686767] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:25:29.688 [2024-11-06 09:01:19.686872] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:29.688 [2024-11-06 09:01:19.686872] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:25:30.260 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:30.260 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@864 -- # return 0 00:25:30.260 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:25:30.260 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@730 -- # xtrace_disable 00:25:30.260 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@78 -- # adq_configure_nvmf_target 0 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # rpc_cmd sock_get_default_impl 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # jq -r .impl_name 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # socket_impl=posix 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@43 -- # rpc_cmd sock_impl_set_options --enable-placement-id 0 --enable-zerocopy-send-server -i posix 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@44 -- # rpc_cmd framework_start_init 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 0 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:30.521 [2024-11-06 09:01:20.525589] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@46 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:30.521 Malloc1 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:30.521 [2024-11-06 09:01:20.576189] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@82 -- # perfpid=1948236 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@83 -- # sleep 2 00:25:30.521 09:01:20 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:25:33.062 09:01:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@85 -- # rpc_cmd nvmf_get_stats 00:25:33.062 09:01:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:33.062 09:01:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:33.063 09:01:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:33.063 09:01:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@85 -- # nvmf_stats='{ 00:25:33.063 "tick_rate": 2400000000, 00:25:33.063 "poll_groups": [ 00:25:33.063 { 00:25:33.063 "name": "nvmf_tgt_poll_group_000", 00:25:33.063 "admin_qpairs": 1, 00:25:33.063 "io_qpairs": 1, 00:25:33.063 "current_admin_qpairs": 1, 00:25:33.063 "current_io_qpairs": 1, 00:25:33.063 "pending_bdev_io": 0, 00:25:33.063 "completed_nvme_io": 19735, 00:25:33.063 "transports": [ 00:25:33.063 { 00:25:33.063 "trtype": "TCP" 00:25:33.063 } 00:25:33.063 ] 00:25:33.063 }, 00:25:33.063 { 00:25:33.063 "name": "nvmf_tgt_poll_group_001", 00:25:33.063 "admin_qpairs": 0, 00:25:33.063 "io_qpairs": 1, 00:25:33.063 "current_admin_qpairs": 0, 00:25:33.063 "current_io_qpairs": 1, 00:25:33.063 "pending_bdev_io": 0, 00:25:33.063 "completed_nvme_io": 27736, 00:25:33.063 "transports": [ 00:25:33.063 { 00:25:33.063 "trtype": "TCP" 00:25:33.063 } 00:25:33.063 ] 00:25:33.063 }, 00:25:33.063 { 00:25:33.063 "name": "nvmf_tgt_poll_group_002", 00:25:33.063 "admin_qpairs": 0, 00:25:33.063 "io_qpairs": 1, 00:25:33.063 "current_admin_qpairs": 0, 00:25:33.063 "current_io_qpairs": 1, 00:25:33.063 "pending_bdev_io": 0, 00:25:33.063 "completed_nvme_io": 24834, 00:25:33.063 "transports": [ 00:25:33.063 { 00:25:33.063 "trtype": "TCP" 00:25:33.063 } 00:25:33.063 ] 00:25:33.063 }, 00:25:33.063 { 00:25:33.063 "name": "nvmf_tgt_poll_group_003", 00:25:33.063 "admin_qpairs": 0, 00:25:33.063 "io_qpairs": 1, 00:25:33.063 "current_admin_qpairs": 0, 00:25:33.063 "current_io_qpairs": 1, 00:25:33.063 "pending_bdev_io": 0, 00:25:33.063 "completed_nvme_io": 20077, 00:25:33.063 "transports": [ 00:25:33.063 { 00:25:33.063 "trtype": "TCP" 00:25:33.063 } 00:25:33.063 ] 00:25:33.063 } 00:25:33.063 ] 00:25:33.063 }' 00:25:33.063 09:01:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 1) | length' 00:25:33.063 09:01:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # wc -l 00:25:33.063 09:01:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@86 -- # count=4 00:25:33.063 09:01:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@87 -- # [[ 4 -ne 4 ]] 00:25:33.063 09:01:22 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@91 -- # wait 1948236 00:25:41.194 Initializing NVMe Controllers 00:25:41.194 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:25:41.194 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:25:41.194 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:25:41.194 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:25:41.194 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:25:41.194 Initialization complete. Launching workers. 00:25:41.194 ======================================================== 00:25:41.194 Latency(us) 00:25:41.194 Device Information : IOPS MiB/s Average min max 00:25:41.194 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 13596.85 53.11 4707.44 1503.25 9191.26 00:25:41.194 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 14431.44 56.37 4434.59 1289.28 9687.10 00:25:41.194 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 14359.74 56.09 4456.73 1299.59 10952.83 00:25:41.194 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 11051.88 43.17 5802.35 1493.61 44932.20 00:25:41.194 ======================================================== 00:25:41.194 Total : 53439.92 208.75 4792.83 1289.28 44932.20 00:25:41.194 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@92 -- # nvmftestfini 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@514 -- # nvmfcleanup 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@121 -- # sync 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@124 -- # set +e 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@125 -- # for i in {1..20} 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:25:41.194 rmmod nvme_tcp 00:25:41.194 rmmod nvme_fabrics 00:25:41.194 rmmod nvme_keyring 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@128 -- # set -e 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@129 -- # return 0 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@515 -- # '[' -n 1948046 ']' 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@516 -- # killprocess 1948046 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@950 -- # '[' -z 1948046 ']' 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@954 -- # kill -0 1948046 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@955 -- # uname 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1948046 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1948046' 00:25:41.194 killing process with pid 1948046 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@969 -- # kill 1948046 00:25:41.194 09:01:30 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@974 -- # wait 1948046 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@297 -- # iptr 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # iptables-save 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # iptables-restore 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@302 -- # remove_spdk_ns 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:41.194 09:01:31 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:43.105 09:01:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:25:43.105 09:01:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@94 -- # adq_reload_driver 00:25:43.105 09:01:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@58 -- # modprobe -a sch_mqprio 00:25:43.105 09:01:33 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@61 -- # rmmod ice 00:25:45.017 09:01:34 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@62 -- # modprobe ice 00:25:46.926 09:01:36 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@63 -- # sleep 5 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@97 -- # nvmftestinit 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@474 -- # prepare_net_devs 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@436 -- # local -g is_hw=no 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@438 -- # remove_spdk_ns 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@309 -- # xtrace_disable 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # pci_devs=() 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@315 -- # local -a pci_devs 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # pci_net_devs=() 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # pci_drivers=() 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@317 -- # local -A pci_drivers 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # net_devs=() 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@319 -- # local -ga net_devs 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # e810=() 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@320 -- # local -ga e810 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # x722=() 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@321 -- # local -ga x722 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # mlx=() 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@322 -- # local -ga mlx 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:25:52.339 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:52.339 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:25:52.340 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:25:52.340 Found net devices under 0000:4b:00.0: cvl_0_0 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ up == up ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:25:52.340 Found net devices under 0000:4b:00.1: cvl_0_1 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@440 -- # is_hw=yes 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:25:52.340 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:52.340 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.666 ms 00:25:52.340 00:25:52.340 --- 10.0.0.2 ping statistics --- 00:25:52.340 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:52.340 rtt min/avg/max/mdev = 0.666/0.666/0.666/0.000 ms 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:52.340 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:52.340 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.298 ms 00:25:52.340 00:25:52.340 --- 10.0.0.1 ping statistics --- 00:25:52.340 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:52.340 rtt min/avg/max/mdev = 0.298/0.298/0.298/0.000 ms 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@448 -- # return 0 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:25:52.340 09:01:41 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@98 -- # adq_configure_driver 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@22 -- # ip netns exec cvl_0_0_ns_spdk ethtool --offload cvl_0_0 hw-tc-offload on 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@24 -- # ip netns exec cvl_0_0_ns_spdk ethtool --set-priv-flags cvl_0_0 channel-pkt-inspect-optimize off 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@26 -- # sysctl -w net.core.busy_poll=1 00:25:52.340 net.core.busy_poll = 1 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@27 -- # sysctl -w net.core.busy_read=1 00:25:52.340 net.core.busy_read = 1 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@29 -- # tc=/usr/sbin/tc 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@31 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 root mqprio num_tc 2 map 0 1 queues 2@0 2@2 hw 1 mode channel 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 ingress 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@35 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc filter add dev cvl_0_0 protocol ip parent ffff: prio 1 flower dst_ip 10.0.0.2/32 ip_proto tcp dst_port 4420 skip_sw hw_tc 1 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@38 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/nvmf/set_xps_rxqs cvl_0_0 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@99 -- # nvmfappstart -m 0xF --wait-for-rpc 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@724 -- # xtrace_disable 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@507 -- # nvmfpid=1953039 00:25:52.340 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@508 -- # waitforlisten 1953039 00:25:52.341 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:25:52.341 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@831 -- # '[' -z 1953039 ']' 00:25:52.341 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:52.341 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@836 -- # local max_retries=100 00:25:52.341 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:52.341 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:52.341 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@840 -- # xtrace_disable 00:25:52.341 09:01:42 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:52.341 [2024-11-06 09:01:42.342992] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:25:52.341 [2024-11-06 09:01:42.343042] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:52.341 [2024-11-06 09:01:42.420631] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:25:52.602 [2024-11-06 09:01:42.456472] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:52.602 [2024-11-06 09:01:42.456505] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:52.602 [2024-11-06 09:01:42.456513] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:52.602 [2024-11-06 09:01:42.456519] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:52.602 [2024-11-06 09:01:42.456525] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:52.602 [2024-11-06 09:01:42.458031] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:25:52.602 [2024-11-06 09:01:42.458148] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:25:52.602 [2024-11-06 09:01:42.458302] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:25:52.602 [2024-11-06 09:01:42.458303] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@864 -- # return 0 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@730 -- # xtrace_disable 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@100 -- # adq_configure_nvmf_target 1 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # rpc_cmd sock_get_default_impl 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # jq -r .impl_name 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@42 -- # socket_impl=posix 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@43 -- # rpc_cmd sock_impl_set_options --enable-placement-id 1 --enable-zerocopy-send-server -i posix 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@44 -- # rpc_cmd framework_start_init 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:53.174 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 1 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:53.435 [2024-11-06 09:01:43.313473] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@46 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:53.435 Malloc1 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:53.435 [2024-11-06 09:01:43.386183] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@104 -- # perfpid=1953122 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@105 -- # sleep 2 00:25:53.435 09:01:43 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@101 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:25:55.349 09:01:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@107 -- # rpc_cmd nvmf_get_stats 00:25:55.349 09:01:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@561 -- # xtrace_disable 00:25:55.349 09:01:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:25:55.349 09:01:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:25:55.349 09:01:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@107 -- # nvmf_stats='{ 00:25:55.349 "tick_rate": 2400000000, 00:25:55.349 "poll_groups": [ 00:25:55.349 { 00:25:55.349 "name": "nvmf_tgt_poll_group_000", 00:25:55.349 "admin_qpairs": 1, 00:25:55.349 "io_qpairs": 2, 00:25:55.349 "current_admin_qpairs": 1, 00:25:55.349 "current_io_qpairs": 2, 00:25:55.349 "pending_bdev_io": 0, 00:25:55.349 "completed_nvme_io": 29276, 00:25:55.349 "transports": [ 00:25:55.349 { 00:25:55.349 "trtype": "TCP" 00:25:55.349 } 00:25:55.349 ] 00:25:55.349 }, 00:25:55.349 { 00:25:55.349 "name": "nvmf_tgt_poll_group_001", 00:25:55.349 "admin_qpairs": 0, 00:25:55.349 "io_qpairs": 2, 00:25:55.349 "current_admin_qpairs": 0, 00:25:55.349 "current_io_qpairs": 2, 00:25:55.349 "pending_bdev_io": 0, 00:25:55.349 "completed_nvme_io": 38478, 00:25:55.349 "transports": [ 00:25:55.349 { 00:25:55.349 "trtype": "TCP" 00:25:55.349 } 00:25:55.349 ] 00:25:55.349 }, 00:25:55.349 { 00:25:55.349 "name": "nvmf_tgt_poll_group_002", 00:25:55.349 "admin_qpairs": 0, 00:25:55.349 "io_qpairs": 0, 00:25:55.349 "current_admin_qpairs": 0, 00:25:55.349 "current_io_qpairs": 0, 00:25:55.349 "pending_bdev_io": 0, 00:25:55.349 "completed_nvme_io": 0, 00:25:55.349 "transports": [ 00:25:55.349 { 00:25:55.349 "trtype": "TCP" 00:25:55.349 } 00:25:55.349 ] 00:25:55.349 }, 00:25:55.349 { 00:25:55.349 "name": "nvmf_tgt_poll_group_003", 00:25:55.349 "admin_qpairs": 0, 00:25:55.349 "io_qpairs": 0, 00:25:55.349 "current_admin_qpairs": 0, 00:25:55.349 "current_io_qpairs": 0, 00:25:55.349 "pending_bdev_io": 0, 00:25:55.349 "completed_nvme_io": 0, 00:25:55.349 "transports": [ 00:25:55.349 { 00:25:55.349 "trtype": "TCP" 00:25:55.349 } 00:25:55.349 ] 00:25:55.349 } 00:25:55.349 ] 00:25:55.349 }' 00:25:55.349 09:01:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 0) | length' 00:25:55.349 09:01:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # wc -l 00:25:55.610 09:01:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@108 -- # count=2 00:25:55.610 09:01:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@109 -- # [[ 2 -lt 2 ]] 00:25:55.610 09:01:45 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@114 -- # wait 1953122 00:26:03.750 Initializing NVMe Controllers 00:26:03.750 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:26:03.750 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:26:03.750 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:26:03.750 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:26:03.750 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:26:03.750 Initialization complete. Launching workers. 00:26:03.750 ======================================================== 00:26:03.750 Latency(us) 00:26:03.750 Device Information : IOPS MiB/s Average min max 00:26:03.750 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 12122.20 47.35 5300.42 899.35 50183.75 00:26:03.750 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 8206.40 32.06 7804.08 1384.99 50471.82 00:26:03.750 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 10145.50 39.63 6326.91 1164.90 51910.22 00:26:03.750 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 9345.00 36.50 6848.22 1064.22 51538.26 00:26:03.750 ======================================================== 00:26:03.750 Total : 39819.09 155.54 6441.19 899.35 51910.22 00:26:03.750 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@115 -- # nvmftestfini 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@514 -- # nvmfcleanup 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@121 -- # sync 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@124 -- # set +e 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@125 -- # for i in {1..20} 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:26:03.750 rmmod nvme_tcp 00:26:03.750 rmmod nvme_fabrics 00:26:03.750 rmmod nvme_keyring 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@128 -- # set -e 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@129 -- # return 0 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@515 -- # '[' -n 1953039 ']' 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@516 -- # killprocess 1953039 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@950 -- # '[' -z 1953039 ']' 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@954 -- # kill -0 1953039 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@955 -- # uname 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1953039 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1953039' 00:26:03.750 killing process with pid 1953039 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@969 -- # kill 1953039 00:26:03.750 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@974 -- # wait 1953039 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@297 -- # iptr 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # iptables-save 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@789 -- # iptables-restore 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@302 -- # remove_spdk_ns 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:04.011 09:01:53 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:05.924 09:01:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:26:05.924 09:01:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- target/perf_adq.sh@117 -- # trap - SIGINT SIGTERM EXIT 00:26:05.924 00:26:05.924 real 0m52.237s 00:26:05.924 user 2m49.681s 00:26:05.924 sys 0m11.124s 00:26:05.924 09:01:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:05.924 09:01:55 nvmf_tcp.nvmf_target_extra.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:26:05.924 ************************************ 00:26:05.924 END TEST nvmf_perf_adq 00:26:05.924 ************************************ 00:26:05.924 09:01:55 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@65 -- # run_test nvmf_shutdown /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:26:05.924 09:01:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:26:05.924 09:01:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:05.924 09:01:55 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:26:05.924 ************************************ 00:26:05.924 START TEST nvmf_shutdown 00:26:05.924 ************************************ 00:26:05.924 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:26:06.185 * Looking for test storage... 00:26:06.185 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1689 -- # lcov --version 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@336 -- # IFS=.-: 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@336 -- # read -ra ver1 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@337 -- # IFS=.-: 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@337 -- # read -ra ver2 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@338 -- # local 'op=<' 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@340 -- # ver1_l=2 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@341 -- # ver2_l=1 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@344 -- # case "$op" in 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@345 -- # : 1 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@365 -- # decimal 1 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@353 -- # local d=1 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@355 -- # echo 1 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@365 -- # ver1[v]=1 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@366 -- # decimal 2 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@353 -- # local d=2 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@355 -- # echo 2 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@366 -- # ver2[v]=2 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@368 -- # return 0 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:26:06.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:06.185 --rc genhtml_branch_coverage=1 00:26:06.185 --rc genhtml_function_coverage=1 00:26:06.185 --rc genhtml_legend=1 00:26:06.185 --rc geninfo_all_blocks=1 00:26:06.185 --rc geninfo_unexecuted_blocks=1 00:26:06.185 00:26:06.185 ' 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:26:06.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:06.185 --rc genhtml_branch_coverage=1 00:26:06.185 --rc genhtml_function_coverage=1 00:26:06.185 --rc genhtml_legend=1 00:26:06.185 --rc geninfo_all_blocks=1 00:26:06.185 --rc geninfo_unexecuted_blocks=1 00:26:06.185 00:26:06.185 ' 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:26:06.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:06.185 --rc genhtml_branch_coverage=1 00:26:06.185 --rc genhtml_function_coverage=1 00:26:06.185 --rc genhtml_legend=1 00:26:06.185 --rc geninfo_all_blocks=1 00:26:06.185 --rc geninfo_unexecuted_blocks=1 00:26:06.185 00:26:06.185 ' 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:26:06.185 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:06.185 --rc genhtml_branch_coverage=1 00:26:06.185 --rc genhtml_function_coverage=1 00:26:06.185 --rc genhtml_legend=1 00:26:06.185 --rc geninfo_all_blocks=1 00:26:06.185 --rc geninfo_unexecuted_blocks=1 00:26:06.185 00:26:06.185 ' 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@7 -- # uname -s 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@15 -- # shopt -s extglob 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@5 -- # export PATH 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@51 -- # : 0 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:26:06.185 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:26:06.185 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:26:06.186 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:26:06.186 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- nvmf/common.sh@55 -- # have_pci_nics=0 00:26:06.186 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@12 -- # MALLOC_BDEV_SIZE=64 00:26:06.186 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:26:06.186 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@162 -- # run_test nvmf_shutdown_tc1 nvmf_shutdown_tc1 00:26:06.186 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:26:06.186 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:06.186 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:26:06.446 ************************************ 00:26:06.446 START TEST nvmf_shutdown_tc1 00:26:06.446 ************************************ 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@1125 -- # nvmf_shutdown_tc1 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@75 -- # starttarget 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@16 -- # nvmftestinit 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@474 -- # prepare_net_devs 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@436 -- # local -g is_hw=no 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@438 -- # remove_spdk_ns 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@309 -- # xtrace_disable 00:26:06.446 09:01:56 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@315 -- # pci_devs=() 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@315 -- # local -a pci_devs 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@317 -- # pci_drivers=() 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@319 -- # net_devs=() 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@319 -- # local -ga net_devs 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@320 -- # e810=() 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@320 -- # local -ga e810 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@321 -- # x722=() 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@321 -- # local -ga x722 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@322 -- # mlx=() 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@322 -- # local -ga mlx 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:26:14.588 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:26:14.588 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:14.588 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:26:14.589 Found net devices under 0000:4b:00.0: cvl_0_0 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:26:14.589 Found net devices under 0000:4b:00.1: cvl_0_1 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@440 -- # is_hw=yes 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:26:14.589 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:14.589 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.693 ms 00:26:14.589 00:26:14.589 --- 10.0.0.2 ping statistics --- 00:26:14.589 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:14.589 rtt min/avg/max/mdev = 0.693/0.693/0.693/0.000 ms 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:14.589 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:14.589 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.318 ms 00:26:14.589 00:26:14.589 --- 10.0.0.1 ping statistics --- 00:26:14.589 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:14.589 rtt min/avg/max/mdev = 0.318/0.318/0.318/0.000 ms 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@448 -- # return 0 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@724 -- # xtrace_disable 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@507 -- # nvmfpid=1959533 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@508 -- # waitforlisten 1959533 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@831 -- # '[' -z 1959533 ']' 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:14.589 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:14.589 09:02:03 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:14.589 [2024-11-06 09:02:03.844157] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:26:14.589 [2024-11-06 09:02:03.844222] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:14.589 [2024-11-06 09:02:03.944606] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:26:14.590 [2024-11-06 09:02:03.996762] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:14.590 [2024-11-06 09:02:03.996813] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:14.590 [2024-11-06 09:02:03.996822] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:14.590 [2024-11-06 09:02:03.996830] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:14.590 [2024-11-06 09:02:03.996836] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:14.590 [2024-11-06 09:02:03.999178] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:14.590 [2024-11-06 09:02:03.999346] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:26:14.590 [2024-11-06 09:02:03.999512] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:26:14.590 [2024-11-06 09:02:03.999513] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@864 -- # return 0 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@730 -- # xtrace_disable 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:14.590 [2024-11-06 09:02:04.688760] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@724 -- # xtrace_disable 00:26:14.590 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@29 -- # cat 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@36 -- # rpc_cmd 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:14.852 09:02:04 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:14.852 Malloc1 00:26:14.852 [2024-11-06 09:02:04.807037] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:14.852 Malloc2 00:26:14.852 Malloc3 00:26:14.852 Malloc4 00:26:14.852 Malloc5 00:26:15.113 Malloc6 00:26:15.113 Malloc7 00:26:15.113 Malloc8 00:26:15.113 Malloc9 00:26:15.113 Malloc10 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@730 -- # xtrace_disable 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@79 -- # perfpid=1959908 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@80 -- # waitforlisten 1959908 /var/tmp/bdevperf.sock 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@831 -- # '[' -z 1959908 ']' 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json /dev/fd/63 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:15.113 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@78 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@558 -- # config=() 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@558 -- # local subsystem config 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:15.113 { 00:26:15.113 "params": { 00:26:15.113 "name": "Nvme$subsystem", 00:26:15.113 "trtype": "$TEST_TRANSPORT", 00:26:15.113 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:15.113 "adrfam": "ipv4", 00:26:15.113 "trsvcid": "$NVMF_PORT", 00:26:15.113 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:15.113 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:15.113 "hdgst": ${hdgst:-false}, 00:26:15.113 "ddgst": ${ddgst:-false} 00:26:15.113 }, 00:26:15.113 "method": "bdev_nvme_attach_controller" 00:26:15.113 } 00:26:15.113 EOF 00:26:15.113 )") 00:26:15.113 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:15.375 { 00:26:15.375 "params": { 00:26:15.375 "name": "Nvme$subsystem", 00:26:15.375 "trtype": "$TEST_TRANSPORT", 00:26:15.375 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:15.375 "adrfam": "ipv4", 00:26:15.375 "trsvcid": "$NVMF_PORT", 00:26:15.375 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:15.375 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:15.375 "hdgst": ${hdgst:-false}, 00:26:15.375 "ddgst": ${ddgst:-false} 00:26:15.375 }, 00:26:15.375 "method": "bdev_nvme_attach_controller" 00:26:15.375 } 00:26:15.375 EOF 00:26:15.375 )") 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:15.375 { 00:26:15.375 "params": { 00:26:15.375 "name": "Nvme$subsystem", 00:26:15.375 "trtype": "$TEST_TRANSPORT", 00:26:15.375 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:15.375 "adrfam": "ipv4", 00:26:15.375 "trsvcid": "$NVMF_PORT", 00:26:15.375 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:15.375 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:15.375 "hdgst": ${hdgst:-false}, 00:26:15.375 "ddgst": ${ddgst:-false} 00:26:15.375 }, 00:26:15.375 "method": "bdev_nvme_attach_controller" 00:26:15.375 } 00:26:15.375 EOF 00:26:15.375 )") 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:15.375 { 00:26:15.375 "params": { 00:26:15.375 "name": "Nvme$subsystem", 00:26:15.375 "trtype": "$TEST_TRANSPORT", 00:26:15.375 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:15.375 "adrfam": "ipv4", 00:26:15.375 "trsvcid": "$NVMF_PORT", 00:26:15.375 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:15.375 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:15.375 "hdgst": ${hdgst:-false}, 00:26:15.375 "ddgst": ${ddgst:-false} 00:26:15.375 }, 00:26:15.375 "method": "bdev_nvme_attach_controller" 00:26:15.375 } 00:26:15.375 EOF 00:26:15.375 )") 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:15.375 { 00:26:15.375 "params": { 00:26:15.375 "name": "Nvme$subsystem", 00:26:15.375 "trtype": "$TEST_TRANSPORT", 00:26:15.375 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:15.375 "adrfam": "ipv4", 00:26:15.375 "trsvcid": "$NVMF_PORT", 00:26:15.375 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:15.375 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:15.375 "hdgst": ${hdgst:-false}, 00:26:15.375 "ddgst": ${ddgst:-false} 00:26:15.375 }, 00:26:15.375 "method": "bdev_nvme_attach_controller" 00:26:15.375 } 00:26:15.375 EOF 00:26:15.375 )") 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:15.375 { 00:26:15.375 "params": { 00:26:15.375 "name": "Nvme$subsystem", 00:26:15.375 "trtype": "$TEST_TRANSPORT", 00:26:15.375 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:15.375 "adrfam": "ipv4", 00:26:15.375 "trsvcid": "$NVMF_PORT", 00:26:15.375 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:15.375 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:15.375 "hdgst": ${hdgst:-false}, 00:26:15.375 "ddgst": ${ddgst:-false} 00:26:15.375 }, 00:26:15.375 "method": "bdev_nvme_attach_controller" 00:26:15.375 } 00:26:15.375 EOF 00:26:15.375 )") 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:15.375 { 00:26:15.375 "params": { 00:26:15.375 "name": "Nvme$subsystem", 00:26:15.375 "trtype": "$TEST_TRANSPORT", 00:26:15.375 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:15.375 "adrfam": "ipv4", 00:26:15.375 "trsvcid": "$NVMF_PORT", 00:26:15.375 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:15.375 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:15.375 "hdgst": ${hdgst:-false}, 00:26:15.375 "ddgst": ${ddgst:-false} 00:26:15.375 }, 00:26:15.375 "method": "bdev_nvme_attach_controller" 00:26:15.375 } 00:26:15.375 EOF 00:26:15.375 )") 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:15.375 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:15.375 { 00:26:15.375 "params": { 00:26:15.375 "name": "Nvme$subsystem", 00:26:15.375 "trtype": "$TEST_TRANSPORT", 00:26:15.375 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:15.375 "adrfam": "ipv4", 00:26:15.375 "trsvcid": "$NVMF_PORT", 00:26:15.375 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:15.375 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:15.375 "hdgst": ${hdgst:-false}, 00:26:15.375 "ddgst": ${ddgst:-false} 00:26:15.375 }, 00:26:15.375 "method": "bdev_nvme_attach_controller" 00:26:15.375 } 00:26:15.375 EOF 00:26:15.375 )") 00:26:15.375 [2024-11-06 09:02:05.277300] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:26:15.376 [2024-11-06 09:02:05.277356] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:26:15.376 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:15.376 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:15.376 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:15.376 { 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme$subsystem", 00:26:15.376 "trtype": "$TEST_TRANSPORT", 00:26:15.376 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "$NVMF_PORT", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:15.376 "hdgst": ${hdgst:-false}, 00:26:15.376 "ddgst": ${ddgst:-false} 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 } 00:26:15.376 EOF 00:26:15.376 )") 00:26:15.376 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:15.376 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:15.376 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:15.376 { 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme$subsystem", 00:26:15.376 "trtype": "$TEST_TRANSPORT", 00:26:15.376 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "$NVMF_PORT", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:15.376 "hdgst": ${hdgst:-false}, 00:26:15.376 "ddgst": ${ddgst:-false} 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 } 00:26:15.376 EOF 00:26:15.376 )") 00:26:15.376 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:15.376 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # jq . 00:26:15.376 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@583 -- # IFS=, 00:26:15.376 09:02:05 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme1", 00:26:15.376 "trtype": "tcp", 00:26:15.376 "traddr": "10.0.0.2", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "4420", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:15.376 "hdgst": false, 00:26:15.376 "ddgst": false 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 },{ 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme2", 00:26:15.376 "trtype": "tcp", 00:26:15.376 "traddr": "10.0.0.2", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "4420", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:26:15.376 "hdgst": false, 00:26:15.376 "ddgst": false 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 },{ 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme3", 00:26:15.376 "trtype": "tcp", 00:26:15.376 "traddr": "10.0.0.2", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "4420", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:26:15.376 "hdgst": false, 00:26:15.376 "ddgst": false 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 },{ 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme4", 00:26:15.376 "trtype": "tcp", 00:26:15.376 "traddr": "10.0.0.2", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "4420", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:26:15.376 "hdgst": false, 00:26:15.376 "ddgst": false 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 },{ 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme5", 00:26:15.376 "trtype": "tcp", 00:26:15.376 "traddr": "10.0.0.2", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "4420", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:26:15.376 "hdgst": false, 00:26:15.376 "ddgst": false 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 },{ 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme6", 00:26:15.376 "trtype": "tcp", 00:26:15.376 "traddr": "10.0.0.2", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "4420", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:26:15.376 "hdgst": false, 00:26:15.376 "ddgst": false 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 },{ 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme7", 00:26:15.376 "trtype": "tcp", 00:26:15.376 "traddr": "10.0.0.2", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "4420", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:26:15.376 "hdgst": false, 00:26:15.376 "ddgst": false 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 },{ 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme8", 00:26:15.376 "trtype": "tcp", 00:26:15.376 "traddr": "10.0.0.2", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "4420", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:26:15.376 "hdgst": false, 00:26:15.376 "ddgst": false 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 },{ 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme9", 00:26:15.376 "trtype": "tcp", 00:26:15.376 "traddr": "10.0.0.2", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "4420", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:26:15.376 "hdgst": false, 00:26:15.376 "ddgst": false 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 },{ 00:26:15.376 "params": { 00:26:15.376 "name": "Nvme10", 00:26:15.376 "trtype": "tcp", 00:26:15.376 "traddr": "10.0.0.2", 00:26:15.376 "adrfam": "ipv4", 00:26:15.376 "trsvcid": "4420", 00:26:15.376 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:26:15.376 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:26:15.376 "hdgst": false, 00:26:15.376 "ddgst": false 00:26:15.376 }, 00:26:15.376 "method": "bdev_nvme_attach_controller" 00:26:15.376 }' 00:26:15.376 [2024-11-06 09:02:05.350073] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:15.376 [2024-11-06 09:02:05.386168] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:16.759 09:02:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:16.759 09:02:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@864 -- # return 0 00:26:16.759 09:02:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@81 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:26:16.759 09:02:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:16.759 09:02:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:16.759 09:02:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:16.759 09:02:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@84 -- # kill -9 1959908 00:26:16.759 09:02:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@85 -- # rm -f /var/run/spdk_bdev1 00:26:16.759 09:02:06 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@88 -- # sleep 1 00:26:17.700 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh: line 74: 1959908 Killed $rootdir/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json <(gen_nvmf_target_json "${num_subsystems[@]}") 00:26:17.700 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@89 -- # kill -0 1959533 00:26:17.700 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:26:17.700 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@92 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:26:17.700 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@558 -- # config=() 00:26:17.700 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@558 -- # local subsystem config 00:26:17.700 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:17.700 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:17.700 { 00:26:17.700 "params": { 00:26:17.700 "name": "Nvme$subsystem", 00:26:17.700 "trtype": "$TEST_TRANSPORT", 00:26:17.700 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:17.700 "adrfam": "ipv4", 00:26:17.700 "trsvcid": "$NVMF_PORT", 00:26:17.700 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:17.700 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:17.700 "hdgst": ${hdgst:-false}, 00:26:17.700 "ddgst": ${ddgst:-false} 00:26:17.700 }, 00:26:17.700 "method": "bdev_nvme_attach_controller" 00:26:17.700 } 00:26:17.700 EOF 00:26:17.700 )") 00:26:17.700 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:17.961 { 00:26:17.961 "params": { 00:26:17.961 "name": "Nvme$subsystem", 00:26:17.961 "trtype": "$TEST_TRANSPORT", 00:26:17.961 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:17.961 "adrfam": "ipv4", 00:26:17.961 "trsvcid": "$NVMF_PORT", 00:26:17.961 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:17.961 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:17.961 "hdgst": ${hdgst:-false}, 00:26:17.961 "ddgst": ${ddgst:-false} 00:26:17.961 }, 00:26:17.961 "method": "bdev_nvme_attach_controller" 00:26:17.961 } 00:26:17.961 EOF 00:26:17.961 )") 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:17.961 { 00:26:17.961 "params": { 00:26:17.961 "name": "Nvme$subsystem", 00:26:17.961 "trtype": "$TEST_TRANSPORT", 00:26:17.961 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:17.961 "adrfam": "ipv4", 00:26:17.961 "trsvcid": "$NVMF_PORT", 00:26:17.961 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:17.961 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:17.961 "hdgst": ${hdgst:-false}, 00:26:17.961 "ddgst": ${ddgst:-false} 00:26:17.961 }, 00:26:17.961 "method": "bdev_nvme_attach_controller" 00:26:17.961 } 00:26:17.961 EOF 00:26:17.961 )") 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:17.961 { 00:26:17.961 "params": { 00:26:17.961 "name": "Nvme$subsystem", 00:26:17.961 "trtype": "$TEST_TRANSPORT", 00:26:17.961 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:17.961 "adrfam": "ipv4", 00:26:17.961 "trsvcid": "$NVMF_PORT", 00:26:17.961 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:17.961 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:17.961 "hdgst": ${hdgst:-false}, 00:26:17.961 "ddgst": ${ddgst:-false} 00:26:17.961 }, 00:26:17.961 "method": "bdev_nvme_attach_controller" 00:26:17.961 } 00:26:17.961 EOF 00:26:17.961 )") 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:17.961 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:17.961 { 00:26:17.961 "params": { 00:26:17.961 "name": "Nvme$subsystem", 00:26:17.961 "trtype": "$TEST_TRANSPORT", 00:26:17.961 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:17.961 "adrfam": "ipv4", 00:26:17.961 "trsvcid": "$NVMF_PORT", 00:26:17.961 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:17.961 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:17.961 "hdgst": ${hdgst:-false}, 00:26:17.962 "ddgst": ${ddgst:-false} 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 } 00:26:17.962 EOF 00:26:17.962 )") 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:17.962 { 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme$subsystem", 00:26:17.962 "trtype": "$TEST_TRANSPORT", 00:26:17.962 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "$NVMF_PORT", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:17.962 "hdgst": ${hdgst:-false}, 00:26:17.962 "ddgst": ${ddgst:-false} 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 } 00:26:17.962 EOF 00:26:17.962 )") 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:17.962 { 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme$subsystem", 00:26:17.962 "trtype": "$TEST_TRANSPORT", 00:26:17.962 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "$NVMF_PORT", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:17.962 "hdgst": ${hdgst:-false}, 00:26:17.962 "ddgst": ${ddgst:-false} 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 } 00:26:17.962 EOF 00:26:17.962 )") 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:17.962 [2024-11-06 09:02:07.860114] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:26:17.962 [2024-11-06 09:02:07.860186] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1960287 ] 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:17.962 { 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme$subsystem", 00:26:17.962 "trtype": "$TEST_TRANSPORT", 00:26:17.962 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "$NVMF_PORT", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:17.962 "hdgst": ${hdgst:-false}, 00:26:17.962 "ddgst": ${ddgst:-false} 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 } 00:26:17.962 EOF 00:26:17.962 )") 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:17.962 { 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme$subsystem", 00:26:17.962 "trtype": "$TEST_TRANSPORT", 00:26:17.962 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "$NVMF_PORT", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:17.962 "hdgst": ${hdgst:-false}, 00:26:17.962 "ddgst": ${ddgst:-false} 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 } 00:26:17.962 EOF 00:26:17.962 )") 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:17.962 { 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme$subsystem", 00:26:17.962 "trtype": "$TEST_TRANSPORT", 00:26:17.962 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "$NVMF_PORT", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:17.962 "hdgst": ${hdgst:-false}, 00:26:17.962 "ddgst": ${ddgst:-false} 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 } 00:26:17.962 EOF 00:26:17.962 )") 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@580 -- # cat 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@582 -- # jq . 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@583 -- # IFS=, 00:26:17.962 09:02:07 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme1", 00:26:17.962 "trtype": "tcp", 00:26:17.962 "traddr": "10.0.0.2", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "4420", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:17.962 "hdgst": false, 00:26:17.962 "ddgst": false 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 },{ 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme2", 00:26:17.962 "trtype": "tcp", 00:26:17.962 "traddr": "10.0.0.2", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "4420", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:26:17.962 "hdgst": false, 00:26:17.962 "ddgst": false 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 },{ 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme3", 00:26:17.962 "trtype": "tcp", 00:26:17.962 "traddr": "10.0.0.2", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "4420", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:26:17.962 "hdgst": false, 00:26:17.962 "ddgst": false 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 },{ 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme4", 00:26:17.962 "trtype": "tcp", 00:26:17.962 "traddr": "10.0.0.2", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "4420", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:26:17.962 "hdgst": false, 00:26:17.962 "ddgst": false 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 },{ 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme5", 00:26:17.962 "trtype": "tcp", 00:26:17.962 "traddr": "10.0.0.2", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "4420", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:26:17.962 "hdgst": false, 00:26:17.962 "ddgst": false 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 },{ 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme6", 00:26:17.962 "trtype": "tcp", 00:26:17.962 "traddr": "10.0.0.2", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "4420", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:26:17.962 "hdgst": false, 00:26:17.962 "ddgst": false 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 },{ 00:26:17.962 "params": { 00:26:17.962 "name": "Nvme7", 00:26:17.962 "trtype": "tcp", 00:26:17.962 "traddr": "10.0.0.2", 00:26:17.962 "adrfam": "ipv4", 00:26:17.962 "trsvcid": "4420", 00:26:17.962 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:26:17.962 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:26:17.962 "hdgst": false, 00:26:17.962 "ddgst": false 00:26:17.962 }, 00:26:17.962 "method": "bdev_nvme_attach_controller" 00:26:17.962 },{ 00:26:17.963 "params": { 00:26:17.963 "name": "Nvme8", 00:26:17.963 "trtype": "tcp", 00:26:17.963 "traddr": "10.0.0.2", 00:26:17.963 "adrfam": "ipv4", 00:26:17.963 "trsvcid": "4420", 00:26:17.963 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:26:17.963 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:26:17.963 "hdgst": false, 00:26:17.963 "ddgst": false 00:26:17.963 }, 00:26:17.963 "method": "bdev_nvme_attach_controller" 00:26:17.963 },{ 00:26:17.963 "params": { 00:26:17.963 "name": "Nvme9", 00:26:17.963 "trtype": "tcp", 00:26:17.963 "traddr": "10.0.0.2", 00:26:17.963 "adrfam": "ipv4", 00:26:17.963 "trsvcid": "4420", 00:26:17.963 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:26:17.963 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:26:17.963 "hdgst": false, 00:26:17.963 "ddgst": false 00:26:17.963 }, 00:26:17.963 "method": "bdev_nvme_attach_controller" 00:26:17.963 },{ 00:26:17.963 "params": { 00:26:17.963 "name": "Nvme10", 00:26:17.963 "trtype": "tcp", 00:26:17.963 "traddr": "10.0.0.2", 00:26:17.963 "adrfam": "ipv4", 00:26:17.963 "trsvcid": "4420", 00:26:17.963 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:26:17.963 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:26:17.963 "hdgst": false, 00:26:17.963 "ddgst": false 00:26:17.963 }, 00:26:17.963 "method": "bdev_nvme_attach_controller" 00:26:17.963 }' 00:26:17.963 [2024-11-06 09:02:07.932859] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:17.963 [2024-11-06 09:02:07.968535] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:19.346 Running I/O for 1 seconds... 00:26:20.547 1862.00 IOPS, 116.38 MiB/s 00:26:20.547 Latency(us) 00:26:20.547 [2024-11-06T08:02:10.660Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:20.547 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:20.547 Verification LBA range: start 0x0 length 0x400 00:26:20.547 Nvme1n1 : 1.16 221.04 13.82 0.00 0.00 286156.59 20643.84 248162.99 00:26:20.547 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:20.547 Verification LBA range: start 0x0 length 0x400 00:26:20.547 Nvme2n1 : 1.12 228.85 14.30 0.00 0.00 270779.95 19005.44 251658.24 00:26:20.547 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:20.547 Verification LBA range: start 0x0 length 0x400 00:26:20.547 Nvme3n1 : 1.08 237.97 14.87 0.00 0.00 255990.40 17476.27 235929.60 00:26:20.547 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:20.547 Verification LBA range: start 0x0 length 0x400 00:26:20.547 Nvme4n1 : 1.08 236.84 14.80 0.00 0.00 252903.25 14964.05 263891.63 00:26:20.547 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:20.547 Verification LBA range: start 0x0 length 0x400 00:26:20.547 Nvme5n1 : 1.15 222.00 13.88 0.00 0.00 265912.32 18131.63 248162.99 00:26:20.547 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:20.547 Verification LBA range: start 0x0 length 0x400 00:26:20.547 Nvme6n1 : 1.18 273.95 17.12 0.00 0.00 211456.08 3932.16 241172.48 00:26:20.547 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:20.547 Verification LBA range: start 0x0 length 0x400 00:26:20.547 Nvme7n1 : 1.14 229.35 14.33 0.00 0.00 242561.99 24357.55 227191.47 00:26:20.547 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:20.547 Verification LBA range: start 0x0 length 0x400 00:26:20.547 Nvme8n1 : 1.19 269.36 16.84 0.00 0.00 208102.91 16165.55 248162.99 00:26:20.547 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:20.547 Verification LBA range: start 0x0 length 0x400 00:26:20.547 Nvme9n1 : 1.19 271.72 16.98 0.00 0.00 202549.10 1426.77 256901.12 00:26:20.547 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:20.547 Verification LBA range: start 0x0 length 0x400 00:26:20.547 Nvme10n1 : 1.20 265.71 16.61 0.00 0.00 203491.97 4177.92 267386.88 00:26:20.547 [2024-11-06T08:02:10.660Z] =================================================================================================================== 00:26:20.547 [2024-11-06T08:02:10.660Z] Total : 2456.80 153.55 0.00 0.00 236864.09 1426.77 267386.88 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@95 -- # stoptarget 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@46 -- # nvmftestfini 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@514 -- # nvmfcleanup 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@121 -- # sync 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@124 -- # set +e 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@125 -- # for i in {1..20} 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:26:20.807 rmmod nvme_tcp 00:26:20.807 rmmod nvme_fabrics 00:26:20.807 rmmod nvme_keyring 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@128 -- # set -e 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@129 -- # return 0 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@515 -- # '[' -n 1959533 ']' 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@516 -- # killprocess 1959533 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@950 -- # '[' -z 1959533 ']' 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@954 -- # kill -0 1959533 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@955 -- # uname 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1959533 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1959533' 00:26:20.807 killing process with pid 1959533 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@969 -- # kill 1959533 00:26:20.807 09:02:10 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@974 -- # wait 1959533 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@297 -- # iptr 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@789 -- # iptables-save 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@789 -- # iptables-restore 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:21.067 09:02:11 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:26:23.610 00:26:23.610 real 0m16.854s 00:26:23.610 user 0m34.482s 00:26:23.610 sys 0m6.818s 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:23.610 ************************************ 00:26:23.610 END TEST nvmf_shutdown_tc1 00:26:23.610 ************************************ 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@163 -- # run_test nvmf_shutdown_tc2 nvmf_shutdown_tc2 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:26:23.610 ************************************ 00:26:23.610 START TEST nvmf_shutdown_tc2 00:26:23.610 ************************************ 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@1125 -- # nvmf_shutdown_tc2 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@100 -- # starttarget 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@16 -- # nvmftestinit 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@474 -- # prepare_net_devs 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@436 -- # local -g is_hw=no 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@438 -- # remove_spdk_ns 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@309 -- # xtrace_disable 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@315 -- # pci_devs=() 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@315 -- # local -a pci_devs 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@317 -- # pci_drivers=() 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@319 -- # net_devs=() 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@319 -- # local -ga net_devs 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@320 -- # e810=() 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@320 -- # local -ga e810 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@321 -- # x722=() 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@321 -- # local -ga x722 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@322 -- # mlx=() 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@322 -- # local -ga mlx 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:26:23.610 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:26:23.610 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:23.610 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:26:23.611 Found net devices under 0000:4b:00.0: cvl_0_0 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:26:23.611 Found net devices under 0000:4b:00.1: cvl_0_1 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@440 -- # is_hw=yes 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:26:23.611 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:23.611 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.627 ms 00:26:23.611 00:26:23.611 --- 10.0.0.2 ping statistics --- 00:26:23.611 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:23.611 rtt min/avg/max/mdev = 0.627/0.627/0.627/0.000 ms 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:23.611 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:23.611 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.281 ms 00:26:23.611 00:26:23.611 --- 10.0.0.1 ping statistics --- 00:26:23.611 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:23.611 rtt min/avg/max/mdev = 0.281/0.281/0.281/0.000 ms 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@448 -- # return 0 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@724 -- # xtrace_disable 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@507 -- # nvmfpid=1961602 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@508 -- # waitforlisten 1961602 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@831 -- # '[' -z 1961602 ']' 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:23.611 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:23.611 09:02:13 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:23.611 [2024-11-06 09:02:13.642472] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:26:23.611 [2024-11-06 09:02:13.642541] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:23.872 [2024-11-06 09:02:13.738116] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:26:23.872 [2024-11-06 09:02:13.773674] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:23.872 [2024-11-06 09:02:13.773703] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:23.872 [2024-11-06 09:02:13.773709] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:23.872 [2024-11-06 09:02:13.773713] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:23.872 [2024-11-06 09:02:13.773717] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:23.872 [2024-11-06 09:02:13.775130] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:23.872 [2024-11-06 09:02:13.775285] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:26:23.872 [2024-11-06 09:02:13.775436] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:23.872 [2024-11-06 09:02:13.775438] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@864 -- # return 0 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@730 -- # xtrace_disable 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:24.443 [2024-11-06 09:02:14.498986] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@724 -- # xtrace_disable 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:24.443 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:26:24.703 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:24.703 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@29 -- # cat 00:26:24.703 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@36 -- # rpc_cmd 00:26:24.703 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:24.703 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:24.703 Malloc1 00:26:24.703 [2024-11-06 09:02:14.605072] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:24.703 Malloc2 00:26:24.703 Malloc3 00:26:24.703 Malloc4 00:26:24.703 Malloc5 00:26:24.703 Malloc6 00:26:24.703 Malloc7 00:26:24.964 Malloc8 00:26:24.964 Malloc9 00:26:24.964 Malloc10 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@730 -- # xtrace_disable 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@104 -- # perfpid=1961830 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@105 -- # waitforlisten 1961830 /var/tmp/bdevperf.sock 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@831 -- # '[' -z 1961830 ']' 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:24.964 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@103 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@558 -- # config=() 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@558 -- # local subsystem config 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:24.964 09:02:14 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:24.964 { 00:26:24.964 "params": { 00:26:24.964 "name": "Nvme$subsystem", 00:26:24.964 "trtype": "$TEST_TRANSPORT", 00:26:24.964 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:24.964 "adrfam": "ipv4", 00:26:24.964 "trsvcid": "$NVMF_PORT", 00:26:24.964 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:24.964 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:24.964 "hdgst": ${hdgst:-false}, 00:26:24.964 "ddgst": ${ddgst:-false} 00:26:24.964 }, 00:26:24.964 "method": "bdev_nvme_attach_controller" 00:26:24.964 } 00:26:24.964 EOF 00:26:24.964 )") 00:26:24.964 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:26:24.964 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:24.964 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:24.964 { 00:26:24.964 "params": { 00:26:24.964 "name": "Nvme$subsystem", 00:26:24.964 "trtype": "$TEST_TRANSPORT", 00:26:24.964 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:24.964 "adrfam": "ipv4", 00:26:24.964 "trsvcid": "$NVMF_PORT", 00:26:24.964 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:24.964 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:24.964 "hdgst": ${hdgst:-false}, 00:26:24.964 "ddgst": ${ddgst:-false} 00:26:24.964 }, 00:26:24.964 "method": "bdev_nvme_attach_controller" 00:26:24.964 } 00:26:24.965 EOF 00:26:24.965 )") 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:24.965 { 00:26:24.965 "params": { 00:26:24.965 "name": "Nvme$subsystem", 00:26:24.965 "trtype": "$TEST_TRANSPORT", 00:26:24.965 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:24.965 "adrfam": "ipv4", 00:26:24.965 "trsvcid": "$NVMF_PORT", 00:26:24.965 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:24.965 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:24.965 "hdgst": ${hdgst:-false}, 00:26:24.965 "ddgst": ${ddgst:-false} 00:26:24.965 }, 00:26:24.965 "method": "bdev_nvme_attach_controller" 00:26:24.965 } 00:26:24.965 EOF 00:26:24.965 )") 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:24.965 { 00:26:24.965 "params": { 00:26:24.965 "name": "Nvme$subsystem", 00:26:24.965 "trtype": "$TEST_TRANSPORT", 00:26:24.965 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:24.965 "adrfam": "ipv4", 00:26:24.965 "trsvcid": "$NVMF_PORT", 00:26:24.965 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:24.965 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:24.965 "hdgst": ${hdgst:-false}, 00:26:24.965 "ddgst": ${ddgst:-false} 00:26:24.965 }, 00:26:24.965 "method": "bdev_nvme_attach_controller" 00:26:24.965 } 00:26:24.965 EOF 00:26:24.965 )") 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:24.965 { 00:26:24.965 "params": { 00:26:24.965 "name": "Nvme$subsystem", 00:26:24.965 "trtype": "$TEST_TRANSPORT", 00:26:24.965 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:24.965 "adrfam": "ipv4", 00:26:24.965 "trsvcid": "$NVMF_PORT", 00:26:24.965 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:24.965 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:24.965 "hdgst": ${hdgst:-false}, 00:26:24.965 "ddgst": ${ddgst:-false} 00:26:24.965 }, 00:26:24.965 "method": "bdev_nvme_attach_controller" 00:26:24.965 } 00:26:24.965 EOF 00:26:24.965 )") 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:24.965 { 00:26:24.965 "params": { 00:26:24.965 "name": "Nvme$subsystem", 00:26:24.965 "trtype": "$TEST_TRANSPORT", 00:26:24.965 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:24.965 "adrfam": "ipv4", 00:26:24.965 "trsvcid": "$NVMF_PORT", 00:26:24.965 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:24.965 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:24.965 "hdgst": ${hdgst:-false}, 00:26:24.965 "ddgst": ${ddgst:-false} 00:26:24.965 }, 00:26:24.965 "method": "bdev_nvme_attach_controller" 00:26:24.965 } 00:26:24.965 EOF 00:26:24.965 )") 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:26:24.965 [2024-11-06 09:02:15.046295] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:26:24.965 [2024-11-06 09:02:15.046348] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1961830 ] 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:24.965 { 00:26:24.965 "params": { 00:26:24.965 "name": "Nvme$subsystem", 00:26:24.965 "trtype": "$TEST_TRANSPORT", 00:26:24.965 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:24.965 "adrfam": "ipv4", 00:26:24.965 "trsvcid": "$NVMF_PORT", 00:26:24.965 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:24.965 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:24.965 "hdgst": ${hdgst:-false}, 00:26:24.965 "ddgst": ${ddgst:-false} 00:26:24.965 }, 00:26:24.965 "method": "bdev_nvme_attach_controller" 00:26:24.965 } 00:26:24.965 EOF 00:26:24.965 )") 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:24.965 { 00:26:24.965 "params": { 00:26:24.965 "name": "Nvme$subsystem", 00:26:24.965 "trtype": "$TEST_TRANSPORT", 00:26:24.965 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:24.965 "adrfam": "ipv4", 00:26:24.965 "trsvcid": "$NVMF_PORT", 00:26:24.965 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:24.965 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:24.965 "hdgst": ${hdgst:-false}, 00:26:24.965 "ddgst": ${ddgst:-false} 00:26:24.965 }, 00:26:24.965 "method": "bdev_nvme_attach_controller" 00:26:24.965 } 00:26:24.965 EOF 00:26:24.965 )") 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:24.965 { 00:26:24.965 "params": { 00:26:24.965 "name": "Nvme$subsystem", 00:26:24.965 "trtype": "$TEST_TRANSPORT", 00:26:24.965 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:24.965 "adrfam": "ipv4", 00:26:24.965 "trsvcid": "$NVMF_PORT", 00:26:24.965 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:24.965 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:24.965 "hdgst": ${hdgst:-false}, 00:26:24.965 "ddgst": ${ddgst:-false} 00:26:24.965 }, 00:26:24.965 "method": "bdev_nvme_attach_controller" 00:26:24.965 } 00:26:24.965 EOF 00:26:24.965 )") 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:24.965 { 00:26:24.965 "params": { 00:26:24.965 "name": "Nvme$subsystem", 00:26:24.965 "trtype": "$TEST_TRANSPORT", 00:26:24.965 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:24.965 "adrfam": "ipv4", 00:26:24.965 "trsvcid": "$NVMF_PORT", 00:26:24.965 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:24.965 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:24.965 "hdgst": ${hdgst:-false}, 00:26:24.965 "ddgst": ${ddgst:-false} 00:26:24.965 }, 00:26:24.965 "method": "bdev_nvme_attach_controller" 00:26:24.965 } 00:26:24.965 EOF 00:26:24.965 )") 00:26:24.965 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@580 -- # cat 00:26:25.226 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@582 -- # jq . 00:26:25.226 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@583 -- # IFS=, 00:26:25.226 09:02:15 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:26:25.226 "params": { 00:26:25.226 "name": "Nvme1", 00:26:25.226 "trtype": "tcp", 00:26:25.226 "traddr": "10.0.0.2", 00:26:25.226 "adrfam": "ipv4", 00:26:25.226 "trsvcid": "4420", 00:26:25.226 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:25.226 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:25.226 "hdgst": false, 00:26:25.226 "ddgst": false 00:26:25.226 }, 00:26:25.226 "method": "bdev_nvme_attach_controller" 00:26:25.226 },{ 00:26:25.226 "params": { 00:26:25.226 "name": "Nvme2", 00:26:25.226 "trtype": "tcp", 00:26:25.226 "traddr": "10.0.0.2", 00:26:25.226 "adrfam": "ipv4", 00:26:25.226 "trsvcid": "4420", 00:26:25.226 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:26:25.226 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:26:25.226 "hdgst": false, 00:26:25.226 "ddgst": false 00:26:25.226 }, 00:26:25.226 "method": "bdev_nvme_attach_controller" 00:26:25.226 },{ 00:26:25.226 "params": { 00:26:25.226 "name": "Nvme3", 00:26:25.226 "trtype": "tcp", 00:26:25.226 "traddr": "10.0.0.2", 00:26:25.226 "adrfam": "ipv4", 00:26:25.226 "trsvcid": "4420", 00:26:25.226 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:26:25.226 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:26:25.226 "hdgst": false, 00:26:25.226 "ddgst": false 00:26:25.226 }, 00:26:25.226 "method": "bdev_nvme_attach_controller" 00:26:25.226 },{ 00:26:25.226 "params": { 00:26:25.226 "name": "Nvme4", 00:26:25.226 "trtype": "tcp", 00:26:25.226 "traddr": "10.0.0.2", 00:26:25.226 "adrfam": "ipv4", 00:26:25.226 "trsvcid": "4420", 00:26:25.226 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:26:25.226 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:26:25.226 "hdgst": false, 00:26:25.226 "ddgst": false 00:26:25.226 }, 00:26:25.226 "method": "bdev_nvme_attach_controller" 00:26:25.226 },{ 00:26:25.226 "params": { 00:26:25.226 "name": "Nvme5", 00:26:25.226 "trtype": "tcp", 00:26:25.226 "traddr": "10.0.0.2", 00:26:25.226 "adrfam": "ipv4", 00:26:25.226 "trsvcid": "4420", 00:26:25.226 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:26:25.226 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:26:25.226 "hdgst": false, 00:26:25.226 "ddgst": false 00:26:25.226 }, 00:26:25.226 "method": "bdev_nvme_attach_controller" 00:26:25.226 },{ 00:26:25.226 "params": { 00:26:25.226 "name": "Nvme6", 00:26:25.226 "trtype": "tcp", 00:26:25.226 "traddr": "10.0.0.2", 00:26:25.226 "adrfam": "ipv4", 00:26:25.226 "trsvcid": "4420", 00:26:25.226 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:26:25.226 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:26:25.226 "hdgst": false, 00:26:25.226 "ddgst": false 00:26:25.226 }, 00:26:25.226 "method": "bdev_nvme_attach_controller" 00:26:25.226 },{ 00:26:25.226 "params": { 00:26:25.226 "name": "Nvme7", 00:26:25.226 "trtype": "tcp", 00:26:25.226 "traddr": "10.0.0.2", 00:26:25.226 "adrfam": "ipv4", 00:26:25.226 "trsvcid": "4420", 00:26:25.226 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:26:25.226 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:26:25.227 "hdgst": false, 00:26:25.227 "ddgst": false 00:26:25.227 }, 00:26:25.227 "method": "bdev_nvme_attach_controller" 00:26:25.227 },{ 00:26:25.227 "params": { 00:26:25.227 "name": "Nvme8", 00:26:25.227 "trtype": "tcp", 00:26:25.227 "traddr": "10.0.0.2", 00:26:25.227 "adrfam": "ipv4", 00:26:25.227 "trsvcid": "4420", 00:26:25.227 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:26:25.227 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:26:25.227 "hdgst": false, 00:26:25.227 "ddgst": false 00:26:25.227 }, 00:26:25.227 "method": "bdev_nvme_attach_controller" 00:26:25.227 },{ 00:26:25.227 "params": { 00:26:25.227 "name": "Nvme9", 00:26:25.227 "trtype": "tcp", 00:26:25.227 "traddr": "10.0.0.2", 00:26:25.227 "adrfam": "ipv4", 00:26:25.227 "trsvcid": "4420", 00:26:25.227 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:26:25.227 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:26:25.227 "hdgst": false, 00:26:25.227 "ddgst": false 00:26:25.227 }, 00:26:25.227 "method": "bdev_nvme_attach_controller" 00:26:25.227 },{ 00:26:25.227 "params": { 00:26:25.227 "name": "Nvme10", 00:26:25.227 "trtype": "tcp", 00:26:25.227 "traddr": "10.0.0.2", 00:26:25.227 "adrfam": "ipv4", 00:26:25.227 "trsvcid": "4420", 00:26:25.227 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:26:25.227 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:26:25.227 "hdgst": false, 00:26:25.227 "ddgst": false 00:26:25.227 }, 00:26:25.227 "method": "bdev_nvme_attach_controller" 00:26:25.227 }' 00:26:25.227 [2024-11-06 09:02:15.117980] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:25.227 [2024-11-06 09:02:15.154108] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:26.611 Running I/O for 10 seconds... 00:26:26.611 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:26.611 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@864 -- # return 0 00:26:26.611 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@106 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:26:26.611 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:26.611 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@108 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@51 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@55 -- # '[' -z Nvme1n1 ']' 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@58 -- # local ret=1 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@59 -- # local i 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i = 10 )) 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:26.872 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # read_io_count=3 00:26:26.873 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # '[' 3 -ge 100 ']' 00:26:26.873 09:02:16 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@68 -- # sleep 0.25 00:26:27.133 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i-- )) 00:26:27.133 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:26:27.133 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:26:27.133 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:26:27.133 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:27.133 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:27.133 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:27.133 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # read_io_count=67 00:26:27.133 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # '[' 67 -ge 100 ']' 00:26:27.133 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@68 -- # sleep 0.25 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i-- )) 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@61 -- # read_io_count=131 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # '[' 131 -ge 100 ']' 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@65 -- # ret=0 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@66 -- # break 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@70 -- # return 0 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@111 -- # killprocess 1961830 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@950 -- # '[' -z 1961830 ']' 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # kill -0 1961830 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@955 -- # uname 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1961830 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1961830' 00:26:27.394 killing process with pid 1961830 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@969 -- # kill 1961830 00:26:27.394 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@974 -- # wait 1961830 00:26:27.655 2312.00 IOPS, 144.50 MiB/s [2024-11-06T08:02:17.768Z] Received shutdown signal, test time was about 1.023418 seconds 00:26:27.655 00:26:27.655 Latency(us) 00:26:27.655 [2024-11-06T08:02:17.768Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:27.655 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:27.655 Verification LBA range: start 0x0 length 0x400 00:26:27.655 Nvme1n1 : 0.98 261.55 16.35 0.00 0.00 241815.04 23046.83 248162.99 00:26:27.655 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:27.655 Verification LBA range: start 0x0 length 0x400 00:26:27.655 Nvme2n1 : 0.97 262.97 16.44 0.00 0.00 235653.12 19005.44 251658.24 00:26:27.655 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:27.655 Verification LBA range: start 0x0 length 0x400 00:26:27.655 Nvme3n1 : 0.97 264.13 16.51 0.00 0.00 229751.36 12342.61 255153.49 00:26:27.655 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:27.655 Verification LBA range: start 0x0 length 0x400 00:26:27.655 Nvme4n1 : 0.96 273.95 17.12 0.00 0.00 215997.19 4014.08 246415.36 00:26:27.655 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:27.655 Verification LBA range: start 0x0 length 0x400 00:26:27.655 Nvme5n1 : 1.02 250.36 15.65 0.00 0.00 223871.57 20425.39 246415.36 00:26:27.655 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:27.655 Verification LBA range: start 0x0 length 0x400 00:26:27.655 Nvme6n1 : 0.96 200.54 12.53 0.00 0.00 282761.67 41287.68 249910.61 00:26:27.655 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:27.655 Verification LBA range: start 0x0 length 0x400 00:26:27.655 Nvme7n1 : 0.95 202.92 12.68 0.00 0.00 272944.07 21626.88 228939.09 00:26:27.655 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:27.655 Verification LBA range: start 0x0 length 0x400 00:26:27.655 Nvme8n1 : 0.96 265.29 16.58 0.00 0.00 204447.15 20425.39 249910.61 00:26:27.655 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:27.655 Verification LBA range: start 0x0 length 0x400 00:26:27.655 Nvme9n1 : 0.96 200.11 12.51 0.00 0.00 264125.72 20097.71 269134.51 00:26:27.655 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:27.655 Verification LBA range: start 0x0 length 0x400 00:26:27.655 Nvme10n1 : 0.99 198.74 12.42 0.00 0.00 245679.33 9666.56 242920.11 00:26:27.655 [2024-11-06T08:02:17.768Z] =================================================================================================================== 00:26:27.655 [2024-11-06T08:02:17.769Z] Total : 2380.56 148.78 0.00 0.00 238918.15 4014.08 269134.51 00:26:27.656 09:02:17 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@114 -- # sleep 1 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@115 -- # kill -0 1961602 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@117 -- # stoptarget 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@46 -- # nvmftestfini 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@514 -- # nvmfcleanup 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@121 -- # sync 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@124 -- # set +e 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@125 -- # for i in {1..20} 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:26:29.038 rmmod nvme_tcp 00:26:29.038 rmmod nvme_fabrics 00:26:29.038 rmmod nvme_keyring 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@128 -- # set -e 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@129 -- # return 0 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@515 -- # '[' -n 1961602 ']' 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@516 -- # killprocess 1961602 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@950 -- # '[' -z 1961602 ']' 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # kill -0 1961602 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@955 -- # uname 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1961602 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1961602' 00:26:29.038 killing process with pid 1961602 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@969 -- # kill 1961602 00:26:29.038 09:02:18 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@974 -- # wait 1961602 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@297 -- # iptr 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@789 -- # iptables-save 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@789 -- # iptables-restore 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:29.038 09:02:19 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:31.583 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:26:31.583 00:26:31.583 real 0m7.923s 00:26:31.583 user 0m24.158s 00:26:31.583 sys 0m1.252s 00:26:31.583 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:31.584 ************************************ 00:26:31.584 END TEST nvmf_shutdown_tc2 00:26:31.584 ************************************ 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@164 -- # run_test nvmf_shutdown_tc3 nvmf_shutdown_tc3 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:26:31.584 ************************************ 00:26:31.584 START TEST nvmf_shutdown_tc3 00:26:31.584 ************************************ 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@1125 -- # nvmf_shutdown_tc3 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@122 -- # starttarget 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@16 -- # nvmftestinit 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@474 -- # prepare_net_devs 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@436 -- # local -g is_hw=no 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@438 -- # remove_spdk_ns 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@309 -- # xtrace_disable 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@315 -- # pci_devs=() 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@315 -- # local -a pci_devs 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@317 -- # pci_drivers=() 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@319 -- # net_devs=() 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@319 -- # local -ga net_devs 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@320 -- # e810=() 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@320 -- # local -ga e810 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@321 -- # x722=() 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@321 -- # local -ga x722 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@322 -- # mlx=() 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@322 -- # local -ga mlx 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:26:31.584 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:26:31.584 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:26:31.584 Found net devices under 0000:4b:00.0: cvl_0_0 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:26:31.584 Found net devices under 0000:4b:00.1: cvl_0_1 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@440 -- # is_hw=yes 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:26:31.584 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:26:31.585 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:31.585 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.422 ms 00:26:31.585 00:26:31.585 --- 10.0.0.2 ping statistics --- 00:26:31.585 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:31.585 rtt min/avg/max/mdev = 0.422/0.422/0.422/0.000 ms 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:31.585 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:31.585 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.302 ms 00:26:31.585 00:26:31.585 --- 10.0.0.1 ping statistics --- 00:26:31.585 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:31.585 rtt min/avg/max/mdev = 0.302/0.302/0.302/0.000 ms 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@448 -- # return 0 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@724 -- # xtrace_disable 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@507 -- # nvmfpid=1963246 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@508 -- # waitforlisten 1963246 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@831 -- # '[' -z 1963246 ']' 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:31.585 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:31.585 09:02:21 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:31.585 [2024-11-06 09:02:21.621335] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:26:31.585 [2024-11-06 09:02:21.621389] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:31.846 [2024-11-06 09:02:21.712173] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:26:31.846 [2024-11-06 09:02:21.742063] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:31.846 [2024-11-06 09:02:21.742089] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:31.846 [2024-11-06 09:02:21.742094] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:31.846 [2024-11-06 09:02:21.742099] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:31.846 [2024-11-06 09:02:21.742103] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:31.846 [2024-11-06 09:02:21.743284] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:31.846 [2024-11-06 09:02:21.743436] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:26:31.846 [2024-11-06 09:02:21.743591] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:31.846 [2024-11-06 09:02:21.743594] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@864 -- # return 0 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@730 -- # xtrace_disable 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:32.418 [2024-11-06 09:02:22.464550] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@724 -- # xtrace_disable 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@29 -- # cat 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@36 -- # rpc_cmd 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:32.418 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:32.680 Malloc1 00:26:32.680 [2024-11-06 09:02:22.580571] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:32.680 Malloc2 00:26:32.680 Malloc3 00:26:32.680 Malloc4 00:26:32.680 Malloc5 00:26:32.680 Malloc6 00:26:32.680 Malloc7 00:26:32.940 Malloc8 00:26:32.940 Malloc9 00:26:32.940 Malloc10 00:26:32.940 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@730 -- # xtrace_disable 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@126 -- # perfpid=1963628 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@127 -- # waitforlisten 1963628 /var/tmp/bdevperf.sock 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@831 -- # '[' -z 1963628 ']' 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:32.941 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@125 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@558 -- # config=() 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@558 -- # local subsystem config 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:32.941 { 00:26:32.941 "params": { 00:26:32.941 "name": "Nvme$subsystem", 00:26:32.941 "trtype": "$TEST_TRANSPORT", 00:26:32.941 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:32.941 "adrfam": "ipv4", 00:26:32.941 "trsvcid": "$NVMF_PORT", 00:26:32.941 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:32.941 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:32.941 "hdgst": ${hdgst:-false}, 00:26:32.941 "ddgst": ${ddgst:-false} 00:26:32.941 }, 00:26:32.941 "method": "bdev_nvme_attach_controller" 00:26:32.941 } 00:26:32.941 EOF 00:26:32.941 )") 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:32.941 { 00:26:32.941 "params": { 00:26:32.941 "name": "Nvme$subsystem", 00:26:32.941 "trtype": "$TEST_TRANSPORT", 00:26:32.941 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:32.941 "adrfam": "ipv4", 00:26:32.941 "trsvcid": "$NVMF_PORT", 00:26:32.941 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:32.941 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:32.941 "hdgst": ${hdgst:-false}, 00:26:32.941 "ddgst": ${ddgst:-false} 00:26:32.941 }, 00:26:32.941 "method": "bdev_nvme_attach_controller" 00:26:32.941 } 00:26:32.941 EOF 00:26:32.941 )") 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:32.941 { 00:26:32.941 "params": { 00:26:32.941 "name": "Nvme$subsystem", 00:26:32.941 "trtype": "$TEST_TRANSPORT", 00:26:32.941 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:32.941 "adrfam": "ipv4", 00:26:32.941 "trsvcid": "$NVMF_PORT", 00:26:32.941 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:32.941 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:32.941 "hdgst": ${hdgst:-false}, 00:26:32.941 "ddgst": ${ddgst:-false} 00:26:32.941 }, 00:26:32.941 "method": "bdev_nvme_attach_controller" 00:26:32.941 } 00:26:32.941 EOF 00:26:32.941 )") 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:32.941 09:02:22 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:32.941 { 00:26:32.941 "params": { 00:26:32.941 "name": "Nvme$subsystem", 00:26:32.941 "trtype": "$TEST_TRANSPORT", 00:26:32.941 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:32.941 "adrfam": "ipv4", 00:26:32.941 "trsvcid": "$NVMF_PORT", 00:26:32.941 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:32.941 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:32.941 "hdgst": ${hdgst:-false}, 00:26:32.941 "ddgst": ${ddgst:-false} 00:26:32.941 }, 00:26:32.941 "method": "bdev_nvme_attach_controller" 00:26:32.941 } 00:26:32.941 EOF 00:26:32.941 )") 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:32.941 { 00:26:32.941 "params": { 00:26:32.941 "name": "Nvme$subsystem", 00:26:32.941 "trtype": "$TEST_TRANSPORT", 00:26:32.941 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:32.941 "adrfam": "ipv4", 00:26:32.941 "trsvcid": "$NVMF_PORT", 00:26:32.941 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:32.941 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:32.941 "hdgst": ${hdgst:-false}, 00:26:32.941 "ddgst": ${ddgst:-false} 00:26:32.941 }, 00:26:32.941 "method": "bdev_nvme_attach_controller" 00:26:32.941 } 00:26:32.941 EOF 00:26:32.941 )") 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:32.941 { 00:26:32.941 "params": { 00:26:32.941 "name": "Nvme$subsystem", 00:26:32.941 "trtype": "$TEST_TRANSPORT", 00:26:32.941 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:32.941 "adrfam": "ipv4", 00:26:32.941 "trsvcid": "$NVMF_PORT", 00:26:32.941 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:32.941 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:32.941 "hdgst": ${hdgst:-false}, 00:26:32.941 "ddgst": ${ddgst:-false} 00:26:32.941 }, 00:26:32.941 "method": "bdev_nvme_attach_controller" 00:26:32.941 } 00:26:32.941 EOF 00:26:32.941 )") 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:32.941 { 00:26:32.941 "params": { 00:26:32.941 "name": "Nvme$subsystem", 00:26:32.941 "trtype": "$TEST_TRANSPORT", 00:26:32.941 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:32.941 "adrfam": "ipv4", 00:26:32.941 "trsvcid": "$NVMF_PORT", 00:26:32.941 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:32.941 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:32.941 "hdgst": ${hdgst:-false}, 00:26:32.941 "ddgst": ${ddgst:-false} 00:26:32.941 }, 00:26:32.941 "method": "bdev_nvme_attach_controller" 00:26:32.941 } 00:26:32.941 EOF 00:26:32.941 )") 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:32.941 { 00:26:32.941 "params": { 00:26:32.941 "name": "Nvme$subsystem", 00:26:32.941 "trtype": "$TEST_TRANSPORT", 00:26:32.941 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:32.941 "adrfam": "ipv4", 00:26:32.941 "trsvcid": "$NVMF_PORT", 00:26:32.941 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:32.941 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:32.941 "hdgst": ${hdgst:-false}, 00:26:32.941 "ddgst": ${ddgst:-false} 00:26:32.941 }, 00:26:32.941 "method": "bdev_nvme_attach_controller" 00:26:32.941 } 00:26:32.941 EOF 00:26:32.941 )") 00:26:32.941 [2024-11-06 09:02:23.032327] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:26:32.941 [2024-11-06 09:02:23.032395] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1963628 ] 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:32.941 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:32.941 { 00:26:32.941 "params": { 00:26:32.941 "name": "Nvme$subsystem", 00:26:32.941 "trtype": "$TEST_TRANSPORT", 00:26:32.942 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:32.942 "adrfam": "ipv4", 00:26:32.942 "trsvcid": "$NVMF_PORT", 00:26:32.942 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:32.942 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:32.942 "hdgst": ${hdgst:-false}, 00:26:32.942 "ddgst": ${ddgst:-false} 00:26:32.942 }, 00:26:32.942 "method": "bdev_nvme_attach_controller" 00:26:32.942 } 00:26:32.942 EOF 00:26:32.942 )") 00:26:32.942 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:26:32.942 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:26:32.942 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:26:32.942 { 00:26:32.942 "params": { 00:26:32.942 "name": "Nvme$subsystem", 00:26:32.942 "trtype": "$TEST_TRANSPORT", 00:26:32.942 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:32.942 "adrfam": "ipv4", 00:26:32.942 "trsvcid": "$NVMF_PORT", 00:26:32.942 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:32.942 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:32.942 "hdgst": ${hdgst:-false}, 00:26:32.942 "ddgst": ${ddgst:-false} 00:26:32.942 }, 00:26:32.942 "method": "bdev_nvme_attach_controller" 00:26:32.942 } 00:26:32.942 EOF 00:26:32.942 )") 00:26:32.942 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@580 -- # cat 00:26:33.204 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@582 -- # jq . 00:26:33.204 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@583 -- # IFS=, 00:26:33.204 09:02:23 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:26:33.204 "params": { 00:26:33.204 "name": "Nvme1", 00:26:33.204 "trtype": "tcp", 00:26:33.204 "traddr": "10.0.0.2", 00:26:33.204 "adrfam": "ipv4", 00:26:33.204 "trsvcid": "4420", 00:26:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:33.204 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:33.204 "hdgst": false, 00:26:33.204 "ddgst": false 00:26:33.204 }, 00:26:33.204 "method": "bdev_nvme_attach_controller" 00:26:33.204 },{ 00:26:33.204 "params": { 00:26:33.204 "name": "Nvme2", 00:26:33.204 "trtype": "tcp", 00:26:33.204 "traddr": "10.0.0.2", 00:26:33.204 "adrfam": "ipv4", 00:26:33.204 "trsvcid": "4420", 00:26:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:26:33.204 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:26:33.204 "hdgst": false, 00:26:33.204 "ddgst": false 00:26:33.204 }, 00:26:33.204 "method": "bdev_nvme_attach_controller" 00:26:33.204 },{ 00:26:33.204 "params": { 00:26:33.204 "name": "Nvme3", 00:26:33.204 "trtype": "tcp", 00:26:33.204 "traddr": "10.0.0.2", 00:26:33.204 "adrfam": "ipv4", 00:26:33.204 "trsvcid": "4420", 00:26:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:26:33.204 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:26:33.204 "hdgst": false, 00:26:33.204 "ddgst": false 00:26:33.204 }, 00:26:33.204 "method": "bdev_nvme_attach_controller" 00:26:33.204 },{ 00:26:33.204 "params": { 00:26:33.204 "name": "Nvme4", 00:26:33.204 "trtype": "tcp", 00:26:33.204 "traddr": "10.0.0.2", 00:26:33.204 "adrfam": "ipv4", 00:26:33.204 "trsvcid": "4420", 00:26:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:26:33.204 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:26:33.204 "hdgst": false, 00:26:33.204 "ddgst": false 00:26:33.204 }, 00:26:33.204 "method": "bdev_nvme_attach_controller" 00:26:33.204 },{ 00:26:33.204 "params": { 00:26:33.204 "name": "Nvme5", 00:26:33.204 "trtype": "tcp", 00:26:33.204 "traddr": "10.0.0.2", 00:26:33.204 "adrfam": "ipv4", 00:26:33.204 "trsvcid": "4420", 00:26:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:26:33.204 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:26:33.204 "hdgst": false, 00:26:33.204 "ddgst": false 00:26:33.204 }, 00:26:33.204 "method": "bdev_nvme_attach_controller" 00:26:33.204 },{ 00:26:33.204 "params": { 00:26:33.204 "name": "Nvme6", 00:26:33.204 "trtype": "tcp", 00:26:33.204 "traddr": "10.0.0.2", 00:26:33.204 "adrfam": "ipv4", 00:26:33.204 "trsvcid": "4420", 00:26:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:26:33.204 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:26:33.204 "hdgst": false, 00:26:33.204 "ddgst": false 00:26:33.204 }, 00:26:33.204 "method": "bdev_nvme_attach_controller" 00:26:33.204 },{ 00:26:33.204 "params": { 00:26:33.204 "name": "Nvme7", 00:26:33.204 "trtype": "tcp", 00:26:33.204 "traddr": "10.0.0.2", 00:26:33.204 "adrfam": "ipv4", 00:26:33.204 "trsvcid": "4420", 00:26:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:26:33.204 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:26:33.204 "hdgst": false, 00:26:33.204 "ddgst": false 00:26:33.204 }, 00:26:33.204 "method": "bdev_nvme_attach_controller" 00:26:33.204 },{ 00:26:33.204 "params": { 00:26:33.204 "name": "Nvme8", 00:26:33.204 "trtype": "tcp", 00:26:33.204 "traddr": "10.0.0.2", 00:26:33.204 "adrfam": "ipv4", 00:26:33.204 "trsvcid": "4420", 00:26:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:26:33.204 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:26:33.204 "hdgst": false, 00:26:33.204 "ddgst": false 00:26:33.204 }, 00:26:33.204 "method": "bdev_nvme_attach_controller" 00:26:33.204 },{ 00:26:33.204 "params": { 00:26:33.204 "name": "Nvme9", 00:26:33.204 "trtype": "tcp", 00:26:33.204 "traddr": "10.0.0.2", 00:26:33.204 "adrfam": "ipv4", 00:26:33.204 "trsvcid": "4420", 00:26:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:26:33.204 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:26:33.204 "hdgst": false, 00:26:33.204 "ddgst": false 00:26:33.204 }, 00:26:33.204 "method": "bdev_nvme_attach_controller" 00:26:33.204 },{ 00:26:33.204 "params": { 00:26:33.204 "name": "Nvme10", 00:26:33.204 "trtype": "tcp", 00:26:33.204 "traddr": "10.0.0.2", 00:26:33.204 "adrfam": "ipv4", 00:26:33.204 "trsvcid": "4420", 00:26:33.204 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:26:33.204 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:26:33.204 "hdgst": false, 00:26:33.204 "ddgst": false 00:26:33.204 }, 00:26:33.204 "method": "bdev_nvme_attach_controller" 00:26:33.204 }' 00:26:33.204 [2024-11-06 09:02:23.104865] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:33.204 [2024-11-06 09:02:23.140995] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:34.590 Running I/O for 10 seconds... 00:26:34.590 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:34.590 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@864 -- # return 0 00:26:34.590 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@128 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:26:34.590 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:34.590 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@131 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@133 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@51 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@55 -- # '[' -z Nvme1n1 ']' 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@58 -- # local ret=1 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@59 -- # local i 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i = 10 )) 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=3 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 3 -ge 100 ']' 00:26:34.851 09:02:24 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@68 -- # sleep 0.25 00:26:35.112 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i-- )) 00:26:35.112 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:26:35.112 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:26:35.112 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:26:35.112 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:35.112 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:35.112 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:35.112 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=67 00:26:35.112 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 67 -ge 100 ']' 00:26:35.112 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@68 -- # sleep 0.25 00:26:35.372 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i-- )) 00:26:35.372 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # (( i != 0 )) 00:26:35.373 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:26:35.373 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # jq -r '.bdevs[0].num_read_ops' 00:26:35.373 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:35.373 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@61 -- # read_io_count=131 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # '[' 131 -ge 100 ']' 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@65 -- # ret=0 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@66 -- # break 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@70 -- # return 0 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@136 -- # killprocess 1963246 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@950 -- # '[' -z 1963246 ']' 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@954 -- # kill -0 1963246 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@955 -- # uname 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1963246 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1963246' 00:26:35.649 killing process with pid 1963246 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@969 -- # kill 1963246 00:26:35.649 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@974 -- # wait 1963246 00:26:35.649 [2024-11-06 09:02:25.571546] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0a3b0 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.571593] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0a3b0 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575571] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575599] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575606] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575611] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575616] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575621] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575626] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575631] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575636] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575640] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575645] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575650] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575655] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575659] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575664] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575669] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575674] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575678] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575687] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575692] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575697] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575702] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575707] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575712] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.649 [2024-11-06 09:02:25.575716] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575721] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575726] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575731] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575735] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575740] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575745] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575753] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575758] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575762] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575767] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575771] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575776] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575781] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575785] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575790] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575795] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575800] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575804] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575809] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575814] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575820] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575825] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575830] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575834] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575839] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575843] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575848] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575854] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575859] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575863] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575867] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575872] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575876] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575881] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575885] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575890] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575894] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.575899] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf38520 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577861] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577885] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577890] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577896] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577901] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577906] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577911] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577916] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577921] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577935] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577940] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577945] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577950] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577955] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577960] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577964] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577969] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577974] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577978] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577983] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577987] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577992] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.577997] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578002] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578007] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578012] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578016] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578021] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578026] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578030] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578035] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578039] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578044] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578049] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.650 [2024-11-06 09:02:25.578054] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578058] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578065] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578069] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578074] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578079] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578083] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578088] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578093] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578098] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578102] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578107] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578112] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578116] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578121] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578126] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578131] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578135] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578140] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578145] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578149] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578154] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578159] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578163] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578168] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578173] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578177] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578182] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.578187] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ad70 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579113] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579140] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579146] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579151] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579155] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579160] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579165] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579170] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579174] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579179] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579184] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579189] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579194] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579198] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579203] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579208] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579212] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579217] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579222] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579226] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579232] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579236] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579241] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579246] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579251] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579255] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579260] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579264] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579270] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579275] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579280] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579284] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579289] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579294] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579299] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579303] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579308] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579312] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579317] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579321] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579326] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579331] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579336] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579340] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579345] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579349] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.651 [2024-11-06 09:02:25.579354] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579358] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579363] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579367] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579372] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579376] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579382] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579386] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579391] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579397] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579401] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579406] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579410] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579415] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579419] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579424] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579429] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b260 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579918] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579933] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579939] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579944] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579949] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579954] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579959] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579965] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579970] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579974] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579979] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579984] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579989] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579994] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.579998] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580003] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580008] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580013] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580018] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580023] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580035] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580041] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580046] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580050] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580055] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580060] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580065] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580070] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580074] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580079] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580084] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580089] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580094] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580098] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580103] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580108] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580113] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580118] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580123] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580127] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580132] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580137] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580141] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580146] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580151] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580156] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580160] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580166] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580171] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580176] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580180] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580185] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580190] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580195] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580199] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580204] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.652 [2024-11-06 09:02:25.580209] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580214] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580218] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580223] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580227] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580232] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580237] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0b730 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580935] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580949] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580954] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580959] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580964] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580969] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580973] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580978] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580983] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580987] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.580992] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581000] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581005] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581009] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581014] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581019] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581024] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581028] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581033] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581038] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581043] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581048] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581053] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581058] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581062] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581067] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581072] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581077] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581082] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581086] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581091] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581096] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581100] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581105] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581110] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581115] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581119] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581124] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581128] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581134] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581139] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581144] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581149] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581154] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581158] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581163] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581168] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581172] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581177] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581181] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581186] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581190] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581195] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581200] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581204] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581209] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581213] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581218] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581223] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581228] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581233] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581237] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581242] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0bc00 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.581991] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.582006] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.582011] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.653 [2024-11-06 09:02:25.582018] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582023] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582028] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582033] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582038] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582043] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582048] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582052] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582057] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582062] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582066] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582071] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582076] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582080] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582085] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582090] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582095] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582100] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582105] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582109] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582114] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582119] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582123] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582128] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582133] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582138] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582143] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582151] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582156] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582161] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582165] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582170] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582174] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582179] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582184] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582189] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582194] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582198] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582203] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582208] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582212] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582217] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582222] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582227] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582232] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582237] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582242] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582246] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582251] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582255] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0c0d0 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582909] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582923] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582928] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582934] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582939] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582946] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582951] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582956] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582960] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582965] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582970] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582974] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582979] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.654 [2024-11-06 09:02:25.582983] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.582988] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.582993] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.582997] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583002] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583007] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583011] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583016] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583020] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583025] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583030] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583034] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583039] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583044] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583048] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583053] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583057] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583062] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583067] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583073] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583078] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583083] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583088] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583092] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583097] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583101] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583106] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583110] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583115] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583119] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583124] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583128] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583133] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583138] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583142] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583147] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583151] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583155] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583160] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583165] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583170] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583174] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583179] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583184] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583189] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583193] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583199] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.583203] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.589230] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589275] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589292] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589308] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589324] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x101fae0 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.589363] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589382] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589389] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589397] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589412] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589426] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1027bb0 is same with the state(6) to be set 00:26:35.655 [2024-11-06 09:02:25.589454] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589471] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589487] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.655 [2024-11-06 09:02:25.589494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.655 [2024-11-06 09:02:25.589506] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589521] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x146d260 is same with the state(6) to be set 00:26:35.656 [2024-11-06 09:02:25.589544] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589561] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589577] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589592] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589607] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x144d960 is same with the state(6) to be set 00:26:35.656 [2024-11-06 09:02:25.589631] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589648] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589663] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589679] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589693] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1029850 is same with the state(6) to be set 00:26:35.656 [2024-11-06 09:02:25.589715] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589732] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589756] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589771] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589786] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10276d0 is same with the state(6) to be set 00:26:35.656 [2024-11-06 09:02:25.589808] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589825] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589840] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589855] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589869] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1454be0 is same with the state(6) to be set 00:26:35.656 [2024-11-06 09:02:25.589893] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589911] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589926] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589941] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589956] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1029cb0 is same with the state(6) to be set 00:26:35.656 [2024-11-06 09:02:25.589978] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.589987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.589995] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.590004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.590012] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.590019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.590028] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.656 [2024-11-06 09:02:25.590035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.590043] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14553a0 is same with the state(6) to be set 00:26:35.656 [2024-11-06 09:02:25.590457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.656 [2024-11-06 09:02:25.590476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.590491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.656 [2024-11-06 09:02:25.590499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.590509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.656 [2024-11-06 09:02:25.590516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.590525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.656 [2024-11-06 09:02:25.590533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.590543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.656 [2024-11-06 09:02:25.590550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.656 [2024-11-06 09:02:25.590559] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.656 [2024-11-06 09:02:25.590566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.590984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.590993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.591001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.591010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.591017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.591027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.591034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.591043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.591050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.591059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.591068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.591078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.591085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.591095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.591102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.591112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.591119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.591129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.591136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.591145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.657 [2024-11-06 09:02:25.591152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.657 [2024-11-06 09:02:25.591162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591228] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591429] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591445] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.591554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.591581] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:35.658 [2024-11-06 09:02:25.592300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.592321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.592333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.592341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.592350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.592358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.592367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.592375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.592384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.592391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.592401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.592409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.592418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.592426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.592435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.592446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.592455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.592463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.592472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.658 [2024-11-06 09:02:25.592480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.658 [2024-11-06 09:02:25.592489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.592894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.592997] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.659 [2024-11-06 09:02:25.593017] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xf0ca90 is same with the state(6) to be set 00:26:35.659 [2024-11-06 09:02:25.602495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.602528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.602540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.602549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.602560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.602568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.602578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.602586] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.602596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.602604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.602614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.659 [2024-11-06 09:02:25.602621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.659 [2024-11-06 09:02:25.602631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602806] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.602989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.602996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.603006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.603013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.603023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.603030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.603040] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.603047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.603056] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2375840 is same with the state(6) to be set 00:26:35.660 [2024-11-06 09:02:25.603278] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x101fae0 (9): Bad file descriptor 00:26:35.660 [2024-11-06 09:02:25.603328] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.660 [2024-11-06 09:02:25.603339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.603349] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.660 [2024-11-06 09:02:25.603356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.603365] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.660 [2024-11-06 09:02:25.603372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.603381] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:26:35.660 [2024-11-06 09:02:25.603388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.603398] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x146da60 is same with the state(6) to be set 00:26:35.660 [2024-11-06 09:02:25.603416] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1027bb0 (9): Bad file descriptor 00:26:35.660 [2024-11-06 09:02:25.603431] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x146d260 (9): Bad file descriptor 00:26:35.660 [2024-11-06 09:02:25.603444] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x144d960 (9): Bad file descriptor 00:26:35.660 [2024-11-06 09:02:25.603461] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1029850 (9): Bad file descriptor 00:26:35.660 [2024-11-06 09:02:25.603477] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10276d0 (9): Bad file descriptor 00:26:35.660 [2024-11-06 09:02:25.603494] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1454be0 (9): Bad file descriptor 00:26:35.660 [2024-11-06 09:02:25.603511] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1029cb0 (9): Bad file descriptor 00:26:35.660 [2024-11-06 09:02:25.603525] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x14553a0 (9): Bad file descriptor 00:26:35.660 [2024-11-06 09:02:25.604784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.660 [2024-11-06 09:02:25.604799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.660 [2024-11-06 09:02:25.604813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.604823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.604835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.604844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.604855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.604865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.604876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.604885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.604897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.604906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.604918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.604927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.604939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.604948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.604959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.604972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.604983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.604992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:32768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:32896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:33024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605104] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:33152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605392] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.661 [2024-11-06 09:02:25.605427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.661 [2024-11-06 09:02:25.605435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605637] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605773] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605843] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.605910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.662 [2024-11-06 09:02:25.605917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.662 [2024-11-06 09:02:25.606179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606264] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.663 [2024-11-06 09:02:25.606779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.663 [2024-11-06 09:02:25.606789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606906] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.606983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.606993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607084] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.607280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.664 [2024-11-06 09:02:25.607287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.664 [2024-11-06 09:02:25.608622] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:26:35.664 [2024-11-06 09:02:25.611372] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode2, 1] resetting controller 00:26:35.664 [2024-11-06 09:02:25.611723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.664 [2024-11-06 09:02:25.611741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1029cb0 with addr=10.0.0.2, port=4420 00:26:35.664 [2024-11-06 09:02:25.611758] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1029cb0 is same with the state(6) to be set 00:26:35.665 [2024-11-06 09:02:25.612548] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:26:35.665 [2024-11-06 09:02:25.612578] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode8, 1] resetting controller 00:26:35.665 [2024-11-06 09:02:25.612591] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode7, 1] resetting controller 00:26:35.665 [2024-11-06 09:02:25.613143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.665 [2024-11-06 09:02:25.613183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1029850 with addr=10.0.0.2, port=4420 00:26:35.665 [2024-11-06 09:02:25.613194] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1029850 is same with the state(6) to be set 00:26:35.665 [2024-11-06 09:02:25.613211] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1029cb0 (9): Bad file descriptor 00:26:35.665 [2024-11-06 09:02:25.613546] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:26:35.665 [2024-11-06 09:02:25.613592] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:26:35.665 [2024-11-06 09:02:25.613631] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:26:35.665 [2024-11-06 09:02:25.613677] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:26:35.665 [2024-11-06 09:02:25.613991] nvme_tcp.c:1184:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:26:35.665 [2024-11-06 09:02:25.614399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.665 [2024-11-06 09:02:25.614415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x146d260 with addr=10.0.0.2, port=4420 00:26:35.665 [2024-11-06 09:02:25.614424] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x146d260 is same with the state(6) to be set 00:26:35.665 [2024-11-06 09:02:25.614752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.665 [2024-11-06 09:02:25.614763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x101fae0 with addr=10.0.0.2, port=4420 00:26:35.665 [2024-11-06 09:02:25.614771] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x101fae0 is same with the state(6) to be set 00:26:35.665 [2024-11-06 09:02:25.614781] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1029850 (9): Bad file descriptor 00:26:35.665 [2024-11-06 09:02:25.614791] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Ctrlr is in error state 00:26:35.665 [2024-11-06 09:02:25.614798] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] controller reinitialization failed 00:26:35.665 [2024-11-06 09:02:25.614807] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] in failed state. 00:26:35.665 [2024-11-06 09:02:25.614842] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x146da60 (9): Bad file descriptor 00:26:35.665 [2024-11-06 09:02:25.614980] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Resetting controller failed. 00:26:35.665 [2024-11-06 09:02:25.615015] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x146d260 (9): Bad file descriptor 00:26:35.665 [2024-11-06 09:02:25.615025] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x101fae0 (9): Bad file descriptor 00:26:35.665 [2024-11-06 09:02:25.615033] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Ctrlr is in error state 00:26:35.665 [2024-11-06 09:02:25.615040] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] controller reinitialization failed 00:26:35.665 [2024-11-06 09:02:25.615047] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] in failed state. 00:26:35.665 [2024-11-06 09:02:25.615097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615257] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.665 [2024-11-06 09:02:25.615350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.665 [2024-11-06 09:02:25.615358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615707] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.666 [2024-11-06 09:02:25.615943] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.666 [2024-11-06 09:02:25.615953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.615960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.615970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.615977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.615987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.615994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616004] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616011] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616028] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616115] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.616199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.616208] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15549b0 is same with the state(6) to be set 00:26:35.667 [2024-11-06 09:02:25.617505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617586] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617626] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.667 [2024-11-06 09:02:25.617834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.667 [2024-11-06 09:02:25.617841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.617851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.617858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.617867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.617875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.617884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.617892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.617901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.617908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.617918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.617925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.617934] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.617942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.617951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.617958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.617968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.617975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.617985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.617992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618154] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.668 [2024-11-06 09:02:25.618350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.668 [2024-11-06 09:02:25.618359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618446] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618513] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.618621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.618629] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1555c70 is same with the state(6) to be set 00:26:35.669 [2024-11-06 09:02:25.619912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.619925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.619938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.619946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.619956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.619963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.619973] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.619980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.619990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.619998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.669 [2024-11-06 09:02:25.620223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.669 [2024-11-06 09:02:25.620232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620257] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620444] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620640] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.670 [2024-11-06 09:02:25.620682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.670 [2024-11-06 09:02:25.620691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620943] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.620987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.620994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.621003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.621015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.621024] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1556970 is same with the state(6) to be set 00:26:35.671 [2024-11-06 09:02:25.622297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622441] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.671 [2024-11-06 09:02:25.622500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.671 [2024-11-06 09:02:25.622510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622598] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622665] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622672] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622715] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.622984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.622993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.623000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.623010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.623017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.623026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.623034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.623043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.623051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.623061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.623068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.623077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.623085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.672 [2024-11-06 09:02:25.623094] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.672 [2024-11-06 09:02:25.623102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.623411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.623420] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1557c30 is same with the state(6) to be set 00:26:35.673 [2024-11-06 09:02:25.624721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:16512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:17152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:17280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:17408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:17792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:17920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.673 [2024-11-06 09:02:25.624976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:18176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.673 [2024-11-06 09:02:25.624984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.624994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:18432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:18560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:18816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:18944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:19200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:19328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:19456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625166] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:19584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:19712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:19840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:19968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625224] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:20096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:20224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:20352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:20480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:20608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:20736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:20864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:20992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:21120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:21248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:21376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:21504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:21632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:21760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:21888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:22016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:22144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:22272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:22400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:22528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.674 [2024-11-06 09:02:25.625563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.674 [2024-11-06 09:02:25.625573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:22656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:22784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:23424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:24064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:24192 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:24320 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:24448 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.625824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.625832] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x122f000 is same with the state(6) to be set 00:26:35.675 [2024-11-06 09:02:25.627091] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Resetting controller failed. 00:26:35.675 [2024-11-06 09:02:25.627113] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode3, 1] resetting controller 00:26:35.675 [2024-11-06 09:02:25.627126] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode4, 1] resetting controller 00:26:35.675 [2024-11-06 09:02:25.627138] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode5, 1] resetting controller 00:26:35.675 [2024-11-06 09:02:25.627174] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] Ctrlr is in error state 00:26:35.675 [2024-11-06 09:02:25.627181] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] controller reinitialization failed 00:26:35.675 [2024-11-06 09:02:25.627190] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] in failed state. 00:26:35.675 [2024-11-06 09:02:25.627202] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] Ctrlr is in error state 00:26:35.675 [2024-11-06 09:02:25.627209] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] controller reinitialization failed 00:26:35.675 [2024-11-06 09:02:25.627216] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] in failed state. 00:26:35.675 [2024-11-06 09:02:25.627257] bdev_nvme.c:3152:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode7, 1] Unable to perform failover, already in progress. 00:26:35.675 [2024-11-06 09:02:25.627271] bdev_nvme.c:3152:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode8, 1] Unable to perform failover, already in progress. 00:26:35.675 [2024-11-06 09:02:25.627288] bdev_nvme.c:3152:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode10, 1] Unable to perform failover, already in progress. 00:26:35.675 [2024-11-06 09:02:25.627300] bdev_nvme.c:3152:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode6, 1] Unable to perform failover, already in progress. 00:26:35.675 [2024-11-06 09:02:25.627370] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode6, 1] resetting controller 00:26:35.675 [2024-11-06 09:02:25.627383] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode10, 1] resetting controller 00:26:35.675 [2024-11-06 09:02:25.627392] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] Resetting controller failed. 00:26:35.675 [2024-11-06 09:02:25.627400] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] Resetting controller failed. 00:26:35.675 [2024-11-06 09:02:25.627790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.675 [2024-11-06 09:02:25.627818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x10276d0 with addr=10.0.0.2, port=4420 00:26:35.675 [2024-11-06 09:02:25.627826] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x10276d0 is same with the state(6) to be set 00:26:35.675 [2024-11-06 09:02:25.628127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.675 [2024-11-06 09:02:25.628137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x14553a0 with addr=10.0.0.2, port=4420 00:26:35.675 [2024-11-06 09:02:25.628144] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x14553a0 is same with the state(6) to be set 00:26:35.675 [2024-11-06 09:02:25.628432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.675 [2024-11-06 09:02:25.628442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1454be0 with addr=10.0.0.2, port=4420 00:26:35.675 [2024-11-06 09:02:25.628449] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1454be0 is same with the state(6) to be set 00:26:35.675 [2024-11-06 09:02:25.629534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.629545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.629557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.629565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.675 [2024-11-06 09:02:25.629575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.675 [2024-11-06 09:02:25.629582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629973] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.629989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.629997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630040] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.676 [2024-11-06 09:02:25.630176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.676 [2024-11-06 09:02:25.630183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630226] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630277] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630302] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630311] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630389] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630406] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630518] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630612] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:35.677 [2024-11-06 09:02:25.630628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:35.677 [2024-11-06 09:02:25.630636] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x122dd60 is same with the state(6) to be set 00:26:35.677 [2024-11-06 09:02:25.632422] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:26:35.677 [2024-11-06 09:02:25.632445] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode2, 1] resetting controller 00:26:35.677 task offset: 27904 on job bdev=Nvme1n1 fails 00:26:35.677 00:26:35.677 Latency(us) 00:26:35.677 [2024-11-06T08:02:25.790Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:35.677 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:35.677 Job: Nvme1n1 ended in about 0.96 seconds with error 00:26:35.677 Verification LBA range: start 0x0 length 0x400 00:26:35.677 Nvme1n1 : 0.96 200.53 12.53 66.84 0.00 236764.05 13434.88 242920.11 00:26:35.677 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:35.677 Job: Nvme2n1 ended in about 0.96 seconds with error 00:26:35.677 Verification LBA range: start 0x0 length 0x400 00:26:35.677 Nvme2n1 : 0.96 203.62 12.73 66.49 0.00 229653.48 14964.05 220200.96 00:26:35.677 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:35.677 Job: Nvme3n1 ended in about 0.97 seconds with error 00:26:35.677 Verification LBA range: start 0x0 length 0x400 00:26:35.677 Nvme3n1 : 0.97 197.92 12.37 65.97 0.00 230409.60 19333.12 263891.63 00:26:35.677 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:35.677 Job: Nvme4n1 ended in about 0.97 seconds with error 00:26:35.677 Verification LBA range: start 0x0 length 0x400 00:26:35.677 Nvme4n1 : 0.97 197.43 12.34 65.81 0.00 226277.33 18131.63 239424.85 00:26:35.677 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:35.677 Job: Nvme5n1 ended in about 0.97 seconds with error 00:26:35.677 Verification LBA range: start 0x0 length 0x400 00:26:35.677 Nvme5n1 : 0.97 131.29 8.21 65.65 0.00 296225.56 15400.96 249910.61 00:26:35.677 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:35.677 Job: Nvme6n1 ended in about 0.98 seconds with error 00:26:35.677 Verification LBA range: start 0x0 length 0x400 00:26:35.677 Nvme6n1 : 0.98 130.97 8.19 65.49 0.00 290700.23 21845.33 262144.00 00:26:35.678 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:35.678 Job: Nvme7n1 ended in about 0.96 seconds with error 00:26:35.678 Verification LBA range: start 0x0 length 0x400 00:26:35.678 Nvme7n1 : 0.96 199.18 12.45 66.39 0.00 209827.20 18022.40 249910.61 00:26:35.678 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:35.678 Job: Nvme8n1 ended in about 0.96 seconds with error 00:26:35.678 Verification LBA range: start 0x0 length 0x400 00:26:35.678 Nvme8n1 : 0.96 199.75 12.48 66.58 0.00 204415.79 18568.53 255153.49 00:26:35.678 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:35.678 Job: Nvme9n1 ended in about 0.98 seconds with error 00:26:35.678 Verification LBA range: start 0x0 length 0x400 00:26:35.678 Nvme9n1 : 0.98 195.02 12.19 65.01 0.00 205509.12 15182.51 241172.48 00:26:35.678 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:26:35.678 Job: Nvme10n1 ended in about 0.98 seconds with error 00:26:35.678 Verification LBA range: start 0x0 length 0x400 00:26:35.678 Nvme10n1 : 0.98 130.65 8.17 65.33 0.00 266185.67 18896.21 270882.13 00:26:35.678 [2024-11-06T08:02:25.791Z] =================================================================================================================== 00:26:35.678 [2024-11-06T08:02:25.791Z] Total : 1786.36 111.65 659.55 0.00 235955.86 13434.88 270882.13 00:26:35.678 [2024-11-06 09:02:25.656958] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:26:35.678 [2024-11-06 09:02:25.657008] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode9, 1] resetting controller 00:26:35.678 [2024-11-06 09:02:25.657466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.678 [2024-11-06 09:02:25.657486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x144d960 with addr=10.0.0.2, port=4420 00:26:35.678 [2024-11-06 09:02:25.657497] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x144d960 is same with the state(6) to be set 00:26:35.678 [2024-11-06 09:02:25.657982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.678 [2024-11-06 09:02:25.658021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1027bb0 with addr=10.0.0.2, port=4420 00:26:35.678 [2024-11-06 09:02:25.658032] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1027bb0 is same with the state(6) to be set 00:26:35.678 [2024-11-06 09:02:25.658050] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x10276d0 (9): Bad file descriptor 00:26:35.678 [2024-11-06 09:02:25.658063] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x14553a0 (9): Bad file descriptor 00:26:35.678 [2024-11-06 09:02:25.658072] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1454be0 (9): Bad file descriptor 00:26:35.678 [2024-11-06 09:02:25.658506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.678 [2024-11-06 09:02:25.658523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1029cb0 with addr=10.0.0.2, port=4420 00:26:35.678 [2024-11-06 09:02:25.658531] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1029cb0 is same with the state(6) to be set 00:26:35.678 [2024-11-06 09:02:25.658756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.678 [2024-11-06 09:02:25.658768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1029850 with addr=10.0.0.2, port=4420 00:26:35.678 [2024-11-06 09:02:25.658776] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1029850 is same with the state(6) to be set 00:26:35.678 [2024-11-06 09:02:25.659095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.678 [2024-11-06 09:02:25.659106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x146da60 with addr=10.0.0.2, port=4420 00:26:35.678 [2024-11-06 09:02:25.659114] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x146da60 is same with the state(6) to be set 00:26:35.678 [2024-11-06 09:02:25.659128] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x144d960 (9): Bad file descriptor 00:26:35.678 [2024-11-06 09:02:25.659138] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1027bb0 (9): Bad file descriptor 00:26:35.678 [2024-11-06 09:02:25.659147] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] Ctrlr is in error state 00:26:35.678 [2024-11-06 09:02:25.659154] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] controller reinitialization failed 00:26:35.678 [2024-11-06 09:02:25.659163] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] in failed state. 00:26:35.678 [2024-11-06 09:02:25.659177] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] Ctrlr is in error state 00:26:35.678 [2024-11-06 09:02:25.659184] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] controller reinitialization failed 00:26:35.678 [2024-11-06 09:02:25.659191] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] in failed state. 00:26:35.678 [2024-11-06 09:02:25.659201] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] Ctrlr is in error state 00:26:35.678 [2024-11-06 09:02:25.659207] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] controller reinitialization failed 00:26:35.678 [2024-11-06 09:02:25.659214] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] in failed state. 00:26:35.678 [2024-11-06 09:02:25.659257] bdev_nvme.c:3152:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode5, 1] Unable to perform failover, already in progress. 00:26:35.678 [2024-11-06 09:02:25.659271] bdev_nvme.c:3152:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode4, 1] Unable to perform failover, already in progress. 00:26:35.678 [2024-11-06 09:02:25.659283] bdev_nvme.c:3152:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode3, 1] Unable to perform failover, already in progress. 00:26:35.678 [2024-11-06 09:02:25.659294] bdev_nvme.c:3152:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode10, 1] Unable to perform failover, already in progress. 00:26:35.678 [2024-11-06 09:02:25.659305] bdev_nvme.c:3152:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode6, 1] Unable to perform failover, already in progress. 00:26:35.678 [2024-11-06 09:02:25.659639] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] Resetting controller failed. 00:26:35.678 [2024-11-06 09:02:25.659651] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] Resetting controller failed. 00:26:35.678 [2024-11-06 09:02:25.659658] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] Resetting controller failed. 00:26:35.678 [2024-11-06 09:02:25.659674] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1029cb0 (9): Bad file descriptor 00:26:35.678 [2024-11-06 09:02:25.659684] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1029850 (9): Bad file descriptor 00:26:35.678 [2024-11-06 09:02:25.659693] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x146da60 (9): Bad file descriptor 00:26:35.678 [2024-11-06 09:02:25.659702] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] Ctrlr is in error state 00:26:35.678 [2024-11-06 09:02:25.659708] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] controller reinitialization failed 00:26:35.678 [2024-11-06 09:02:25.659715] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] in failed state. 00:26:35.678 [2024-11-06 09:02:25.659725] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] Ctrlr is in error state 00:26:35.678 [2024-11-06 09:02:25.659732] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] controller reinitialization failed 00:26:35.678 [2024-11-06 09:02:25.659738] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] in failed state. 00:26:35.678 [2024-11-06 09:02:25.659977] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode7, 1] resetting controller 00:26:35.678 [2024-11-06 09:02:25.659992] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode8, 1] resetting controller 00:26:35.678 [2024-11-06 09:02:25.660001] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] Resetting controller failed. 00:26:35.678 [2024-11-06 09:02:25.660009] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] Resetting controller failed. 00:26:35.678 [2024-11-06 09:02:25.660028] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Ctrlr is in error state 00:26:35.678 [2024-11-06 09:02:25.660036] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] controller reinitialization failed 00:26:35.678 [2024-11-06 09:02:25.660043] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] in failed state. 00:26:35.678 [2024-11-06 09:02:25.660052] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Ctrlr is in error state 00:26:35.678 [2024-11-06 09:02:25.660059] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] controller reinitialization failed 00:26:35.678 [2024-11-06 09:02:25.660066] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] in failed state. 00:26:35.678 [2024-11-06 09:02:25.660075] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] Ctrlr is in error state 00:26:35.678 [2024-11-06 09:02:25.660082] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] controller reinitialization failed 00:26:35.678 [2024-11-06 09:02:25.660088] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] in failed state. 00:26:35.678 [2024-11-06 09:02:25.660121] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Resetting controller failed. 00:26:35.678 [2024-11-06 09:02:25.660129] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] Resetting controller failed. 00:26:35.678 [2024-11-06 09:02:25.660137] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] Resetting controller failed. 00:26:35.678 [2024-11-06 09:02:25.660455] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.678 [2024-11-06 09:02:25.660469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x101fae0 with addr=10.0.0.2, port=4420 00:26:35.679 [2024-11-06 09:02:25.660476] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x101fae0 is same with the state(6) to be set 00:26:35.679 [2024-11-06 09:02:25.660686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:35.679 [2024-11-06 09:02:25.660697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x146d260 with addr=10.0.0.2, port=4420 00:26:35.679 [2024-11-06 09:02:25.660704] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x146d260 is same with the state(6) to be set 00:26:35.679 [2024-11-06 09:02:25.660733] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x101fae0 (9): Bad file descriptor 00:26:35.679 [2024-11-06 09:02:25.660744] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x146d260 (9): Bad file descriptor 00:26:35.679 [2024-11-06 09:02:25.660775] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] Ctrlr is in error state 00:26:35.679 [2024-11-06 09:02:25.660783] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] controller reinitialization failed 00:26:35.679 [2024-11-06 09:02:25.660790] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] in failed state. 00:26:35.679 [2024-11-06 09:02:25.660799] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] Ctrlr is in error state 00:26:35.679 [2024-11-06 09:02:25.660806] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] controller reinitialization failed 00:26:35.679 [2024-11-06 09:02:25.660816] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] in failed state. 00:26:35.679 [2024-11-06 09:02:25.660843] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] Resetting controller failed. 00:26:35.679 [2024-11-06 09:02:25.660851] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] Resetting controller failed. 00:26:35.940 09:02:25 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@137 -- # sleep 1 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@138 -- # NOT wait 1963628 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@650 -- # local es=0 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@652 -- # valid_exec_arg wait 1963628 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@638 -- # local arg=wait 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@642 -- # type -t wait 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@653 -- # wait 1963628 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@653 -- # es=255 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@662 -- # es=127 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@663 -- # case "$es" in 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@670 -- # es=1 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@140 -- # stoptarget 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@46 -- # nvmftestfini 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@514 -- # nvmfcleanup 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@121 -- # sync 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@124 -- # set +e 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@125 -- # for i in {1..20} 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:26:36.883 rmmod nvme_tcp 00:26:36.883 rmmod nvme_fabrics 00:26:36.883 rmmod nvme_keyring 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@128 -- # set -e 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@129 -- # return 0 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@515 -- # '[' -n 1963246 ']' 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@516 -- # killprocess 1963246 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@950 -- # '[' -z 1963246 ']' 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@954 -- # kill -0 1963246 00:26:36.883 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (1963246) - No such process 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@977 -- # echo 'Process with pid 1963246 is not found' 00:26:36.883 Process with pid 1963246 is not found 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@297 -- # iptr 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@789 -- # iptables-save 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@789 -- # iptables-restore 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:36.883 09:02:26 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:39.427 09:02:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:26:39.427 00:26:39.427 real 0m7.749s 00:26:39.427 user 0m19.295s 00:26:39.427 sys 0m1.168s 00:26:39.427 09:02:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:39.427 09:02:28 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:26:39.427 ************************************ 00:26:39.427 END TEST nvmf_shutdown_tc3 00:26:39.427 ************************************ 00:26:39.427 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@166 -- # [[ e810 == \e\8\1\0 ]] 00:26:39.427 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@166 -- # [[ tcp == \r\d\m\a ]] 00:26:39.427 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@167 -- # run_test nvmf_shutdown_tc4 nvmf_shutdown_tc4 00:26:39.427 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:26:39.427 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:39.427 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:26:39.428 ************************************ 00:26:39.428 START TEST nvmf_shutdown_tc4 00:26:39.428 ************************************ 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@1125 -- # nvmf_shutdown_tc4 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@145 -- # starttarget 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@16 -- # nvmftestinit 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@474 -- # prepare_net_devs 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@436 -- # local -g is_hw=no 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@438 -- # remove_spdk_ns 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@309 -- # xtrace_disable 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@315 -- # pci_devs=() 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@315 -- # local -a pci_devs 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@316 -- # pci_net_devs=() 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@317 -- # pci_drivers=() 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@317 -- # local -A pci_drivers 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@319 -- # net_devs=() 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@319 -- # local -ga net_devs 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@320 -- # e810=() 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@320 -- # local -ga e810 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@321 -- # x722=() 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@321 -- # local -ga x722 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@322 -- # mlx=() 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@322 -- # local -ga mlx 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:26:39.428 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:26:39.428 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:26:39.428 Found net devices under 0000:4b:00.0: cvl_0_0 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:26:39.428 Found net devices under 0000:4b:00.1: cvl_0_1 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@440 -- # is_hw=yes 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:26:39.428 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:26:39.429 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:39.429 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.621 ms 00:26:39.429 00:26:39.429 --- 10.0.0.2 ping statistics --- 00:26:39.429 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:39.429 rtt min/avg/max/mdev = 0.621/0.621/0.621/0.000 ms 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:39.429 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:39.429 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.205 ms 00:26:39.429 00:26:39.429 --- 10.0.0.1 ping statistics --- 00:26:39.429 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:39.429 rtt min/avg/max/mdev = 0.205/0.205/0.205/0.000 ms 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@448 -- # return 0 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@19 -- # nvmfappstart -m 0x1E 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@724 -- # xtrace_disable 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@507 -- # nvmfpid=1964997 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@508 -- # waitforlisten 1964997 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@831 -- # '[' -z 1964997 ']' 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:39.429 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:39.429 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:26:39.429 [2024-11-06 09:02:29.472151] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:26:39.429 [2024-11-06 09:02:29.472213] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:39.690 [2024-11-06 09:02:29.563212] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:26:39.690 [2024-11-06 09:02:29.593411] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:39.690 [2024-11-06 09:02:29.593441] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:39.690 [2024-11-06 09:02:29.593447] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:39.690 [2024-11-06 09:02:29.593452] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:39.690 [2024-11-06 09:02:29.593456] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:39.690 [2024-11-06 09:02:29.594936] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:39.690 [2024-11-06 09:02:29.595185] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:26:39.690 [2024-11-06 09:02:29.595342] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:39.690 [2024-11-06 09:02:29.595344] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@864 -- # return 0 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@730 -- # xtrace_disable 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:26:39.690 [2024-11-06 09:02:29.730787] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@23 -- # num_subsystems=({1..10}) 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@25 -- # timing_enter create_subsystems 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@724 -- # xtrace_disable 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@27 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@28 -- # for i in "${num_subsystems[@]}" 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@29 -- # cat 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@36 -- # rpc_cmd 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:39.690 09:02:29 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:26:39.951 Malloc1 00:26:39.951 [2024-11-06 09:02:29.841605] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:39.951 Malloc2 00:26:39.951 Malloc3 00:26:39.951 Malloc4 00:26:39.951 Malloc5 00:26:39.951 Malloc6 00:26:39.951 Malloc7 00:26:40.211 Malloc8 00:26:40.211 Malloc9 00:26:40.211 Malloc10 00:26:40.211 09:02:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:40.211 09:02:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@37 -- # timing_exit create_subsystems 00:26:40.211 09:02:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@730 -- # xtrace_disable 00:26:40.211 09:02:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:26:40.211 09:02:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@149 -- # perfpid=1965144 00:26:40.211 09:02:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@150 -- # sleep 5 00:26:40.211 09:02:30 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@148 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 45056 -O 4096 -w randwrite -t 20 -r 'trtype:tcp adrfam:IPV4 traddr:10.0.0.2 trsvcid:4420' -P 4 00:26:40.211 [2024-11-06 09:02:30.306651] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@152 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@155 -- # killprocess 1964997 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@950 -- # '[' -z 1964997 ']' 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@954 -- # kill -0 1964997 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@955 -- # uname 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1964997 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1964997' 00:26:45.617 killing process with pid 1964997 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@969 -- # kill 1964997 00:26:45.617 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@974 -- # wait 1964997 00:26:45.617 [2024-11-06 09:02:35.322958] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9440 is same with the state(6) to be set 00:26:45.617 [2024-11-06 09:02:35.323009] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9440 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.323016] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9440 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.323024] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9440 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.323030] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9440 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.323611] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9910 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.323636] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9910 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.323642] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9910 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324004] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9de0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324028] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9de0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324034] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9de0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324040] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9de0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324045] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9de0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324050] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9de0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324055] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9de0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324071] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9de0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324076] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed9de0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324295] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f70 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324319] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f70 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324326] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f70 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324331] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f70 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324336] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f70 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324342] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f70 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.324347] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8f70 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325138] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8720 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325157] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8720 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325163] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8720 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325168] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8720 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325173] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8720 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325178] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8720 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325448] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8bf0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325464] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8bf0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325469] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8bf0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325474] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8bf0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325479] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8bf0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.325484] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xed8bf0 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.328643] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdab9e0 is same with the state(6) to be set 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 [2024-11-06 09:02:35.328897] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaab10 is same with the state(6) to be set 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 [2024-11-06 09:02:35.328917] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaab10 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.328923] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaab10 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.328928] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaab10 is same with the state(6) to be set 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 [2024-11-06 09:02:35.328939] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaab10 is same with the state(6) to be set 00:26:45.618 [2024-11-06 09:02:35.328944] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaab10 is same with the state(6) to be set 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 [2024-11-06 09:02:35.329533] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 Write completed with error (sct=0, sc=8) 00:26:45.618 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 [2024-11-06 09:02:35.330477] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:26:45.619 starting I/O failed: -6 00:26:45.619 starting I/O failed: -6 00:26:45.619 starting I/O failed: -6 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 [2024-11-06 09:02:35.331577] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 [2024-11-06 09:02:35.331697] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xedb120 is same with tstarting I/O failed: -6 00:26:45.619 he state(6) to be set 00:26:45.619 [2024-11-06 09:02:35.331714] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xedb120 is same with tWrite completed with error (sct=0, sc=8) 00:26:45.619 he state(6) to be set 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 [2024-11-06 09:02:35.332107] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeda2b0 is same with tWrite completed with error (sct=0, sc=8) 00:26:45.619 he state(6) to be set 00:26:45.619 starting I/O failed: -6 00:26:45.619 [2024-11-06 09:02:35.332123] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeda2b0 is same with the state(6) to be set 00:26:45.619 [2024-11-06 09:02:35.332129] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeda2b0 is same with the state(6) to be set 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 [2024-11-06 09:02:35.332133] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeda2b0 is same with the state(6) to be set 00:26:45.619 starting I/O failed: -6 00:26:45.619 [2024-11-06 09:02:35.332139] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeda2b0 is same with the state(6) to be set 00:26:45.619 [2024-11-06 09:02:35.332144] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xeda2b0 is same with the state(6) to be set 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.619 starting I/O failed: -6 00:26:45.619 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 [2024-11-06 09:02:35.332817] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode3, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:26:45.620 [2024-11-06 09:02:35.332820] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaa150 is same with the state(6) to be set 00:26:45.620 NVMe io qpair process completion error 00:26:45.620 [2024-11-06 09:02:35.332836] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaa150 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.332842] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaa150 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.332846] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaa150 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.332851] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaa150 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.332856] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaa150 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333033] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaa620 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333047] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaa620 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333052] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaa620 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333057] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdaa620 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333779] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333800] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333805] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333810] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333815] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333819] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333824] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333829] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333834] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333839] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333843] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333848] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45f70 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333902] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdabed0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333916] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdabed0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333925] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdabed0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333930] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdabed0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333934] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdabed0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333939] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdabed0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333944] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdabed0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333948] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdabed0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333953] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdabed0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.333958] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdabed0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334217] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdac3a0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334231] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xdac3a0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334410] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334425] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334430] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334436] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334440] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334445] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334451] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334456] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334460] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334465] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 [2024-11-06 09:02:35.334469] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe45aa0 is same with the state(6) to be set 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 [2024-11-06 09:02:35.335200] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode8, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:45.620 starting I/O failed: -6 00:26:45.620 starting I/O failed: -6 00:26:45.620 starting I/O failed: -6 00:26:45.620 NVMe io qpair process completion error 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 starting I/O failed: -6 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.620 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 [2024-11-06 09:02:35.337080] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:26:45.621 starting I/O failed: -6 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 [2024-11-06 09:02:35.338044] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 [2024-11-06 09:02:35.338964] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.621 Write completed with error (sct=0, sc=8) 00:26:45.621 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 [2024-11-06 09:02:35.340411] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode2, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:26:45.622 NVMe io qpair process completion error 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 [2024-11-06 09:02:35.341736] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.622 starting I/O failed: -6 00:26:45.622 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 [2024-11-06 09:02:35.342539] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 [2024-11-06 09:02:35.343451] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 Write completed with error (sct=0, sc=8) 00:26:45.623 starting I/O failed: -6 00:26:45.623 [2024-11-06 09:02:35.346241] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode9, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:26:45.624 NVMe io qpair process completion error 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 [2024-11-06 09:02:35.347464] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 [2024-11-06 09:02:35.348277] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 [2024-11-06 09:02:35.349249] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.624 starting I/O failed: -6 00:26:45.624 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 [2024-11-06 09:02:35.350877] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode4, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:26:45.625 NVMe io qpair process completion error 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 [2024-11-06 09:02:35.352237] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 [2024-11-06 09:02:35.353054] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.625 starting I/O failed: -6 00:26:45.625 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 [2024-11-06 09:02:35.353993] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 [2024-11-06 09:02:35.356578] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode5, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:45.626 NVMe io qpair process completion error 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 Write completed with error (sct=0, sc=8) 00:26:45.626 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 [2024-11-06 09:02:35.357808] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:26:45.627 starting I/O failed: -6 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 [2024-11-06 09:02:35.358762] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 [2024-11-06 09:02:35.359683] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.627 Write completed with error (sct=0, sc=8) 00:26:45.627 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 [2024-11-06 09:02:35.361321] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode10, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:26:45.628 NVMe io qpair process completion error 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 [2024-11-06 09:02:35.362518] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 Write completed with error (sct=0, sc=8) 00:26:45.628 starting I/O failed: -6 00:26:45.628 [2024-11-06 09:02:35.363350] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 [2024-11-06 09:02:35.364317] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.629 Write completed with error (sct=0, sc=8) 00:26:45.629 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 [2024-11-06 09:02:35.366260] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:45.630 NVMe io qpair process completion error 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 [2024-11-06 09:02:35.367488] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:26:45.630 starting I/O failed: -6 00:26:45.630 starting I/O failed: -6 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 [2024-11-06 09:02:35.368417] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.630 Write completed with error (sct=0, sc=8) 00:26:45.630 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 [2024-11-06 09:02:35.369319] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 [2024-11-06 09:02:35.372638] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode6, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:45.631 NVMe io qpair process completion error 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 [2024-11-06 09:02:35.373687] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.631 Write completed with error (sct=0, sc=8) 00:26:45.631 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 [2024-11-06 09:02:35.374579] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 [2024-11-06 09:02:35.375483] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.632 starting I/O failed: -6 00:26:45.632 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 starting I/O failed: -6 00:26:45.633 [2024-11-06 09:02:35.377147] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode7, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:26:45.633 NVMe io qpair process completion error 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.633 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Write completed with error (sct=0, sc=8) 00:26:45.634 Initializing NVMe Controllers 00:26:45.634 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode7 00:26:45.634 Controller IO queue size 128, less than required. 00:26:45.634 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:26:45.634 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode3 00:26:45.634 Controller IO queue size 128, less than required. 00:26:45.634 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:26:45.634 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode8 00:26:45.634 Controller IO queue size 128, less than required. 00:26:45.634 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:26:45.634 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode2 00:26:45.634 Controller IO queue size 128, less than required. 00:26:45.634 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:26:45.634 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode9 00:26:45.634 Controller IO queue size 128, less than required. 00:26:45.634 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:26:45.634 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode4 00:26:45.634 Controller IO queue size 128, less than required. 00:26:45.634 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:26:45.634 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode5 00:26:45.634 Controller IO queue size 128, less than required. 00:26:45.634 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:26:45.634 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode10 00:26:45.634 Controller IO queue size 128, less than required. 00:26:45.634 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:26:45.634 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:26:45.634 Controller IO queue size 128, less than required. 00:26:45.634 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:26:45.634 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode6 00:26:45.634 Controller IO queue size 128, less than required. 00:26:45.634 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:26:45.634 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode7) NSID 1 with lcore 0 00:26:45.634 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode3) NSID 1 with lcore 0 00:26:45.634 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode8) NSID 1 with lcore 0 00:26:45.634 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode2) NSID 1 with lcore 0 00:26:45.634 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode9) NSID 1 with lcore 0 00:26:45.634 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode4) NSID 1 with lcore 0 00:26:45.634 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode5) NSID 1 with lcore 0 00:26:45.634 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode10) NSID 1 with lcore 0 00:26:45.634 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:26:45.634 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode6) NSID 1 with lcore 0 00:26:45.634 Initialization complete. Launching workers. 00:26:45.634 ======================================================== 00:26:45.634 Latency(us) 00:26:45.634 Device Information : IOPS MiB/s Average min max 00:26:45.634 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode7) NSID 1 from core 0: 1849.07 79.45 69240.57 935.17 119832.61 00:26:45.634 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode3) NSID 1 from core 0: 1864.74 80.13 67975.22 680.82 149309.50 00:26:45.634 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode8) NSID 1 from core 0: 1898.19 81.56 67348.82 572.68 148996.29 00:26:45.634 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode2) NSID 1 from core 0: 1889.93 81.21 67087.49 606.07 118104.96 00:26:45.634 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode9) NSID 1 from core 0: 1902.21 81.74 66674.77 837.54 116252.11 00:26:45.634 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode4) NSID 1 from core 0: 1913.86 82.24 66306.61 845.82 120078.98 00:26:45.634 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode5) NSID 1 from core 0: 1914.07 82.25 66323.30 607.33 118248.78 00:26:45.634 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode10) NSID 1 from core 0: 1885.06 81.00 67379.95 672.22 123706.55 00:26:45.634 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1900.73 81.67 66846.58 785.67 125170.07 00:26:45.634 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode6) NSID 1 from core 0: 1906.24 81.91 66680.51 713.70 127565.73 00:26:45.634 ======================================================== 00:26:45.634 Total : 18924.10 813.14 67177.74 572.68 149309.50 00:26:45.634 00:26:45.634 [2024-11-06 09:02:35.386324] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1210410 is same with the state(6) to be set 00:26:45.634 [2024-11-06 09:02:35.386371] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x120f560 is same with the state(6) to be set 00:26:45.634 [2024-11-06 09:02:35.386401] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1210740 is same with the state(6) to be set 00:26:45.634 [2024-11-06 09:02:35.386430] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1211ae0 is same with the state(6) to be set 00:26:45.634 [2024-11-06 09:02:35.386459] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1210a70 is same with the state(6) to be set 00:26:45.634 [2024-11-06 09:02:35.386489] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x120f890 is same with the state(6) to be set 00:26:45.634 [2024-11-06 09:02:35.386518] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x120fbc0 is same with the state(6) to be set 00:26:45.634 [2024-11-06 09:02:35.386547] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1211900 is same with the state(6) to be set 00:26:45.634 [2024-11-06 09:02:35.386576] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1211720 is same with the state(6) to be set 00:26:45.634 [2024-11-06 09:02:35.386604] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x120fef0 is same with the state(6) to be set 00:26:45.634 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:26:45.634 09:02:35 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@156 -- # sleep 1 00:26:46.577 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@158 -- # NOT wait 1965144 00:26:46.577 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@650 -- # local es=0 00:26:46.577 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@652 -- # valid_exec_arg wait 1965144 00:26:46.577 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@638 -- # local arg=wait 00:26:46.577 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@642 -- # type -t wait 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@653 -- # wait 1965144 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@653 -- # es=1 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@159 -- # stoptarget 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@42 -- # rm -f ./local-job0-0-verify.state 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@44 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- target/shutdown.sh@46 -- # nvmftestfini 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@514 -- # nvmfcleanup 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@121 -- # sync 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@124 -- # set +e 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@125 -- # for i in {1..20} 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:26:46.578 rmmod nvme_tcp 00:26:46.578 rmmod nvme_fabrics 00:26:46.578 rmmod nvme_keyring 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@128 -- # set -e 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@129 -- # return 0 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@515 -- # '[' -n 1964997 ']' 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@516 -- # killprocess 1964997 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@950 -- # '[' -z 1964997 ']' 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@954 -- # kill -0 1964997 00:26:46.578 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (1964997) - No such process 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@977 -- # echo 'Process with pid 1964997 is not found' 00:26:46.578 Process with pid 1964997 is not found 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@297 -- # iptr 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@789 -- # iptables-save 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@789 -- # iptables-restore 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@302 -- # remove_spdk_ns 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:46.578 09:02:36 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:49.123 09:02:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:26:49.123 00:26:49.123 real 0m9.660s 00:26:49.123 user 0m25.611s 00:26:49.123 sys 0m3.902s 00:26:49.123 09:02:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:49.123 09:02:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown.nvmf_shutdown_tc4 -- common/autotest_common.sh@10 -- # set +x 00:26:49.123 ************************************ 00:26:49.123 END TEST nvmf_shutdown_tc4 00:26:49.123 ************************************ 00:26:49.123 09:02:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- target/shutdown.sh@170 -- # trap - SIGINT SIGTERM EXIT 00:26:49.123 00:26:49.123 real 0m42.746s 00:26:49.123 user 1m43.823s 00:26:49.123 sys 0m13.453s 00:26:49.123 09:02:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:49.123 09:02:38 nvmf_tcp.nvmf_target_extra.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:26:49.123 ************************************ 00:26:49.123 END TEST nvmf_shutdown 00:26:49.123 ************************************ 00:26:49.123 09:02:38 nvmf_tcp.nvmf_target_extra -- nvmf/nvmf_target_extra.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:26:49.123 00:26:49.123 real 12m50.505s 00:26:49.123 user 27m30.951s 00:26:49.123 sys 3m43.826s 00:26:49.123 09:02:38 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@1126 -- # xtrace_disable 00:26:49.123 09:02:38 nvmf_tcp.nvmf_target_extra -- common/autotest_common.sh@10 -- # set +x 00:26:49.123 ************************************ 00:26:49.123 END TEST nvmf_target_extra 00:26:49.123 ************************************ 00:26:49.123 09:02:38 nvmf_tcp -- nvmf/nvmf.sh@16 -- # run_test nvmf_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_host.sh --transport=tcp 00:26:49.123 09:02:38 nvmf_tcp -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:26:49.123 09:02:38 nvmf_tcp -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:49.123 09:02:38 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:26:49.123 ************************************ 00:26:49.123 START TEST nvmf_host 00:26:49.123 ************************************ 00:26:49.123 09:02:38 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_host.sh --transport=tcp 00:26:49.123 * Looking for test storage... 00:26:49.123 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:26:49.123 09:02:38 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:26:49.123 09:02:38 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1689 -- # lcov --version 00:26:49.123 09:02:38 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@336 -- # IFS=.-: 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@336 -- # read -ra ver1 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@337 -- # IFS=.-: 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@337 -- # read -ra ver2 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@338 -- # local 'op=<' 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@340 -- # ver1_l=2 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@341 -- # ver2_l=1 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@344 -- # case "$op" in 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@345 -- # : 1 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@365 -- # decimal 1 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@353 -- # local d=1 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@355 -- # echo 1 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@365 -- # ver1[v]=1 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@366 -- # decimal 2 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@353 -- # local d=2 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@355 -- # echo 2 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@366 -- # ver2[v]=2 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@368 -- # return 0 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:26:49.123 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:49.123 --rc genhtml_branch_coverage=1 00:26:49.123 --rc genhtml_function_coverage=1 00:26:49.123 --rc genhtml_legend=1 00:26:49.123 --rc geninfo_all_blocks=1 00:26:49.123 --rc geninfo_unexecuted_blocks=1 00:26:49.123 00:26:49.123 ' 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:26:49.123 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:49.123 --rc genhtml_branch_coverage=1 00:26:49.123 --rc genhtml_function_coverage=1 00:26:49.123 --rc genhtml_legend=1 00:26:49.123 --rc geninfo_all_blocks=1 00:26:49.123 --rc geninfo_unexecuted_blocks=1 00:26:49.123 00:26:49.123 ' 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:26:49.123 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:49.123 --rc genhtml_branch_coverage=1 00:26:49.123 --rc genhtml_function_coverage=1 00:26:49.123 --rc genhtml_legend=1 00:26:49.123 --rc geninfo_all_blocks=1 00:26:49.123 --rc geninfo_unexecuted_blocks=1 00:26:49.123 00:26:49.123 ' 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:26:49.123 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:49.123 --rc genhtml_branch_coverage=1 00:26:49.123 --rc genhtml_function_coverage=1 00:26:49.123 --rc genhtml_legend=1 00:26:49.123 --rc geninfo_all_blocks=1 00:26:49.123 --rc geninfo_unexecuted_blocks=1 00:26:49.123 00:26:49.123 ' 00:26:49.123 09:02:39 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@7 -- # uname -s 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@15 -- # shopt -s extglob 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- paths/export.sh@5 -- # export PATH 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@51 -- # : 0 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:26:49.124 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@11 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@13 -- # TEST_ARGS=("$@") 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@15 -- # [[ 0 -eq 0 ]] 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@16 -- # run_test nvmf_multicontroller /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:26:49.124 ************************************ 00:26:49.124 START TEST nvmf_multicontroller 00:26:49.124 ************************************ 00:26:49.124 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:26:49.386 * Looking for test storage... 00:26:49.386 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1689 -- # lcov --version 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@333 -- # local ver1 ver1_l 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@334 -- # local ver2 ver2_l 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@336 -- # IFS=.-: 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@336 -- # read -ra ver1 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@337 -- # IFS=.-: 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@337 -- # read -ra ver2 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@338 -- # local 'op=<' 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@340 -- # ver1_l=2 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@341 -- # ver2_l=1 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@344 -- # case "$op" in 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@345 -- # : 1 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@364 -- # (( v = 0 )) 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@365 -- # decimal 1 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@353 -- # local d=1 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@355 -- # echo 1 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@365 -- # ver1[v]=1 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@366 -- # decimal 2 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@353 -- # local d=2 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@355 -- # echo 2 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@366 -- # ver2[v]=2 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@368 -- # return 0 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:26:49.386 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:49.386 --rc genhtml_branch_coverage=1 00:26:49.386 --rc genhtml_function_coverage=1 00:26:49.386 --rc genhtml_legend=1 00:26:49.386 --rc geninfo_all_blocks=1 00:26:49.386 --rc geninfo_unexecuted_blocks=1 00:26:49.386 00:26:49.386 ' 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:26:49.386 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:49.386 --rc genhtml_branch_coverage=1 00:26:49.386 --rc genhtml_function_coverage=1 00:26:49.386 --rc genhtml_legend=1 00:26:49.386 --rc geninfo_all_blocks=1 00:26:49.386 --rc geninfo_unexecuted_blocks=1 00:26:49.386 00:26:49.386 ' 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:26:49.386 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:49.386 --rc genhtml_branch_coverage=1 00:26:49.386 --rc genhtml_function_coverage=1 00:26:49.386 --rc genhtml_legend=1 00:26:49.386 --rc geninfo_all_blocks=1 00:26:49.386 --rc geninfo_unexecuted_blocks=1 00:26:49.386 00:26:49.386 ' 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:26:49.386 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:26:49.386 --rc genhtml_branch_coverage=1 00:26:49.386 --rc genhtml_function_coverage=1 00:26:49.386 --rc genhtml_legend=1 00:26:49.386 --rc geninfo_all_blocks=1 00:26:49.386 --rc geninfo_unexecuted_blocks=1 00:26:49.386 00:26:49.386 ' 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@7 -- # uname -s 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@15 -- # shopt -s extglob 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:49.386 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@5 -- # export PATH 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@51 -- # : 0 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:26:49.387 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@55 -- # have_pci_nics=0 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@11 -- # MALLOC_BDEV_SIZE=64 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@13 -- # NVMF_HOST_FIRST_PORT=60000 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@14 -- # NVMF_HOST_SECOND_PORT=60001 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@18 -- # '[' tcp == rdma ']' 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@23 -- # nvmftestinit 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@474 -- # prepare_net_devs 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@436 -- # local -g is_hw=no 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@438 -- # remove_spdk_ns 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@309 -- # xtrace_disable 00:26:49.387 09:02:39 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@315 -- # pci_devs=() 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@315 -- # local -a pci_devs 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@316 -- # pci_net_devs=() 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@317 -- # pci_drivers=() 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@317 -- # local -A pci_drivers 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@319 -- # net_devs=() 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@319 -- # local -ga net_devs 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@320 -- # e810=() 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@320 -- # local -ga e810 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@321 -- # x722=() 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@321 -- # local -ga x722 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@322 -- # mlx=() 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@322 -- # local -ga mlx 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:26:55.972 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:26:55.972 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:26:55.973 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:26:55.973 Found net devices under 0000:4b:00.0: cvl_0_0 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@416 -- # [[ up == up ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:26:55.973 Found net devices under 0000:4b:00.1: cvl_0_1 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@440 -- # is_hw=yes 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:26:55.973 09:02:45 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:56.234 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:56.234 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:56.234 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:26:56.234 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:26:56.235 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:56.235 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.677 ms 00:26:56.235 00:26:56.235 --- 10.0.0.2 ping statistics --- 00:26:56.235 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:56.235 rtt min/avg/max/mdev = 0.677/0.677/0.677/0.000 ms 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:56.235 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:56.235 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.287 ms 00:26:56.235 00:26:56.235 --- 10.0.0.1 ping statistics --- 00:26:56.235 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:56.235 rtt min/avg/max/mdev = 0.287/0.287/0.287/0.000 ms 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@448 -- # return 0 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@25 -- # nvmfappstart -m 0xE 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@724 -- # xtrace_disable 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@507 -- # nvmfpid=1970557 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@508 -- # waitforlisten 1970557 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@831 -- # '[' -z 1970557 ']' 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:56.235 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:56.235 09:02:46 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:56.235 [2024-11-06 09:02:46.345949] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:26:56.235 [2024-11-06 09:02:46.346013] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:56.496 [2024-11-06 09:02:46.441486] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:26:56.496 [2024-11-06 09:02:46.475423] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:56.496 [2024-11-06 09:02:46.475457] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:56.496 [2024-11-06 09:02:46.475463] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:56.496 [2024-11-06 09:02:46.475468] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:56.496 [2024-11-06 09:02:46.475472] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:56.496 [2024-11-06 09:02:46.476644] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:26:56.496 [2024-11-06 09:02:46.476801] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:26:56.496 [2024-11-06 09:02:46.476803] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:26:57.068 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:57.068 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@864 -- # return 0 00:26:57.068 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:26:57.068 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@730 -- # xtrace_disable 00:26:57.068 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.329 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:57.329 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@27 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:26:57.329 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.329 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.329 [2024-11-06 09:02:47.188429] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@29 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.330 Malloc0 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@30 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.330 [2024-11-06 09:02:47.255485] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.330 [2024-11-06 09:02:47.267425] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.330 Malloc1 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc1 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4421 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@44 -- # bdevperf_pid=1970906 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@46 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; pap "$testdir/try.txt"; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w write -t 1 -f 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@47 -- # waitforlisten 1970906 /var/tmp/bdevperf.sock 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@831 -- # '[' -z 1970906 ']' 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@836 -- # local max_retries=100 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:26:57.330 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@840 -- # xtrace_disable 00:26:57.330 09:02:47 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@864 -- # return 0 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@50 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.270 NVMe0n1 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@54 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@54 -- # grep -c NVMe 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:58.270 1 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@60 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:26:58.270 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@650 -- # local es=0 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -q nqn.2021-09-7.io.spdk:00001 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.271 request: 00:26:58.271 { 00:26:58.271 "name": "NVMe0", 00:26:58.271 "trtype": "tcp", 00:26:58.271 "traddr": "10.0.0.2", 00:26:58.271 "adrfam": "ipv4", 00:26:58.271 "trsvcid": "4420", 00:26:58.271 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:58.271 "hostnqn": "nqn.2021-09-7.io.spdk:00001", 00:26:58.271 "hostaddr": "10.0.0.1", 00:26:58.271 "prchk_reftag": false, 00:26:58.271 "prchk_guard": false, 00:26:58.271 "hdgst": false, 00:26:58.271 "ddgst": false, 00:26:58.271 "allow_unrecognized_csi": false, 00:26:58.271 "method": "bdev_nvme_attach_controller", 00:26:58.271 "req_id": 1 00:26:58.271 } 00:26:58.271 Got JSON-RPC error response 00:26:58.271 response: 00:26:58.271 { 00:26:58.271 "code": -114, 00:26:58.271 "message": "A controller named NVMe0 already exists with the specified network path" 00:26:58.271 } 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # es=1 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@65 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@650 -- # local es=0 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.1 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.271 request: 00:26:58.271 { 00:26:58.271 "name": "NVMe0", 00:26:58.271 "trtype": "tcp", 00:26:58.271 "traddr": "10.0.0.2", 00:26:58.271 "adrfam": "ipv4", 00:26:58.271 "trsvcid": "4420", 00:26:58.271 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:26:58.271 "hostaddr": "10.0.0.1", 00:26:58.271 "prchk_reftag": false, 00:26:58.271 "prchk_guard": false, 00:26:58.271 "hdgst": false, 00:26:58.271 "ddgst": false, 00:26:58.271 "allow_unrecognized_csi": false, 00:26:58.271 "method": "bdev_nvme_attach_controller", 00:26:58.271 "req_id": 1 00:26:58.271 } 00:26:58.271 Got JSON-RPC error response 00:26:58.271 response: 00:26:58.271 { 00:26:58.271 "code": -114, 00:26:58.271 "message": "A controller named NVMe0 already exists with the specified network path" 00:26:58.271 } 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # es=1 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@69 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@650 -- # local es=0 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x disable 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.271 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.271 request: 00:26:58.271 { 00:26:58.271 "name": "NVMe0", 00:26:58.271 "trtype": "tcp", 00:26:58.271 "traddr": "10.0.0.2", 00:26:58.533 "adrfam": "ipv4", 00:26:58.533 "trsvcid": "4420", 00:26:58.533 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:58.533 "hostaddr": "10.0.0.1", 00:26:58.533 "prchk_reftag": false, 00:26:58.533 "prchk_guard": false, 00:26:58.533 "hdgst": false, 00:26:58.533 "ddgst": false, 00:26:58.533 "multipath": "disable", 00:26:58.533 "allow_unrecognized_csi": false, 00:26:58.533 "method": "bdev_nvme_attach_controller", 00:26:58.533 "req_id": 1 00:26:58.533 } 00:26:58.533 Got JSON-RPC error response 00:26:58.533 response: 00:26:58.533 { 00:26:58.533 "code": -114, 00:26:58.533 "message": "A controller named NVMe0 already exists and multipath is disabled" 00:26:58.533 } 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # es=1 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@74 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@650 -- # local es=0 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 -x failover 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.533 request: 00:26:58.533 { 00:26:58.533 "name": "NVMe0", 00:26:58.533 "trtype": "tcp", 00:26:58.533 "traddr": "10.0.0.2", 00:26:58.533 "adrfam": "ipv4", 00:26:58.533 "trsvcid": "4420", 00:26:58.533 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:58.533 "hostaddr": "10.0.0.1", 00:26:58.533 "prchk_reftag": false, 00:26:58.533 "prchk_guard": false, 00:26:58.533 "hdgst": false, 00:26:58.533 "ddgst": false, 00:26:58.533 "multipath": "failover", 00:26:58.533 "allow_unrecognized_csi": false, 00:26:58.533 "method": "bdev_nvme_attach_controller", 00:26:58.533 "req_id": 1 00:26:58.533 } 00:26:58.533 Got JSON-RPC error response 00:26:58.533 response: 00:26:58.533 { 00:26:58.533 "code": -114, 00:26:58.533 "message": "A controller named NVMe0 already exists with the specified network path" 00:26:58.533 } 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@653 -- # es=1 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@79 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.533 NVMe0n1 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@83 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@87 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe1 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.1 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.533 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.796 00:26:58.796 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:58.796 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:26:58.796 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # grep -c NVMe 00:26:58.796 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:58.796 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:58.796 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:58.796 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@90 -- # '[' 2 '!=' 2 ']' 00:26:58.796 09:02:48 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:26:59.736 { 00:26:59.736 "results": [ 00:26:59.736 { 00:26:59.736 "job": "NVMe0n1", 00:26:59.736 "core_mask": "0x1", 00:26:59.736 "workload": "write", 00:26:59.736 "status": "finished", 00:26:59.736 "queue_depth": 128, 00:26:59.736 "io_size": 4096, 00:26:59.736 "runtime": 1.007712, 00:26:59.736 "iops": 27241.910387094726, 00:26:59.736 "mibps": 106.41371244958877, 00:26:59.736 "io_failed": 0, 00:26:59.736 "io_timeout": 0, 00:26:59.736 "avg_latency_us": 4689.108319976687, 00:26:59.736 "min_latency_us": 2048.0, 00:26:59.736 "max_latency_us": 10922.666666666666 00:26:59.736 } 00:26:59.736 ], 00:26:59.736 "core_count": 1 00:26:59.736 } 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@98 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe1 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@100 -- # [[ -n '' ]] 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@116 -- # killprocess 1970906 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@950 -- # '[' -z 1970906 ']' 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@954 -- # kill -0 1970906 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@955 -- # uname 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:26:59.736 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1970906 00:26:59.996 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:26:59.996 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:26:59.996 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1970906' 00:26:59.996 killing process with pid 1970906 00:26:59.996 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@969 -- # kill 1970906 00:26:59.996 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@974 -- # wait 1970906 00:26:59.996 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@118 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:26:59.996 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.996 09:02:49 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@119 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@561 -- # xtrace_disable 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@121 -- # trap - SIGINT SIGTERM EXIT 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@123 -- # pap /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1595 -- # read -r file 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1594 -- # find /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt -type f 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1594 -- # sort -u 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1596 -- # cat 00:26:59.996 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:26:59.996 [2024-11-06 09:02:47.394691] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:26:59.996 [2024-11-06 09:02:47.394758] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1970906 ] 00:26:59.996 [2024-11-06 09:02:47.465918] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:59.996 [2024-11-06 09:02:47.502060] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:26:59.996 [2024-11-06 09:02:48.644599] bdev.c:4897:bdev_name_add: *ERROR*: Bdev name afe721fe-fb4a-4523-9b24-de8521e4c725 already exists 00:26:59.996 [2024-11-06 09:02:48.644631] bdev.c:8100:bdev_register: *ERROR*: Unable to add uuid:afe721fe-fb4a-4523-9b24-de8521e4c725 alias for bdev NVMe1n1 00:26:59.996 [2024-11-06 09:02:48.644640] bdev_nvme.c:4604:nvme_bdev_create: *ERROR*: spdk_bdev_register() failed 00:26:59.996 Running I/O for 1 seconds... 00:26:59.996 27197.00 IOPS, 106.24 MiB/s 00:26:59.996 Latency(us) 00:26:59.996 [2024-11-06T08:02:50.109Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:59.996 Job: NVMe0n1 (Core Mask 0x1, workload: write, depth: 128, IO size: 4096) 00:26:59.996 NVMe0n1 : 1.01 27241.91 106.41 0.00 0.00 4689.11 2048.00 10922.67 00:26:59.996 [2024-11-06T08:02:50.109Z] =================================================================================================================== 00:26:59.996 [2024-11-06T08:02:50.109Z] Total : 27241.91 106.41 0.00 0.00 4689.11 2048.00 10922.67 00:26:59.996 Received shutdown signal, test time was about 1.000000 seconds 00:26:59.996 00:26:59.996 Latency(us) 00:26:59.996 [2024-11-06T08:02:50.109Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:59.996 [2024-11-06T08:02:50.109Z] =================================================================================================================== 00:26:59.996 [2024-11-06T08:02:50.109Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:59.996 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1601 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1595 -- # read -r file 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- host/multicontroller.sh@124 -- # nvmftestfini 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@514 -- # nvmfcleanup 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@121 -- # sync 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@124 -- # set +e 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@125 -- # for i in {1..20} 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:26:59.996 rmmod nvme_tcp 00:26:59.996 rmmod nvme_fabrics 00:26:59.996 rmmod nvme_keyring 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@128 -- # set -e 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@129 -- # return 0 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@515 -- # '[' -n 1970557 ']' 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@516 -- # killprocess 1970557 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@950 -- # '[' -z 1970557 ']' 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@954 -- # kill -0 1970557 00:26:59.996 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@955 -- # uname 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1970557 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1970557' 00:27:00.258 killing process with pid 1970557 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@969 -- # kill 1970557 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@974 -- # wait 1970557 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@297 -- # iptr 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@789 -- # iptables-save 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@789 -- # iptables-restore 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:00.258 09:02:50 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_multicontroller -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:02.804 00:27:02.804 real 0m13.227s 00:27:02.804 user 0m16.456s 00:27:02.804 sys 0m5.969s 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@1126 -- # xtrace_disable 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:27:02.804 ************************************ 00:27:02.804 END TEST nvmf_multicontroller 00:27:02.804 ************************************ 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@17 -- # run_test nvmf_aer /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:27:02.804 ************************************ 00:27:02.804 START TEST nvmf_aer 00:27:02.804 ************************************ 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:27:02.804 * Looking for test storage... 00:27:02.804 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1689 -- # lcov --version 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@336 -- # IFS=.-: 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@336 -- # read -ra ver1 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@337 -- # IFS=.-: 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@337 -- # read -ra ver2 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@338 -- # local 'op=<' 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@340 -- # ver1_l=2 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@341 -- # ver2_l=1 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@344 -- # case "$op" in 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@345 -- # : 1 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@365 -- # decimal 1 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@353 -- # local d=1 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@355 -- # echo 1 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@365 -- # ver1[v]=1 00:27:02.804 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@366 -- # decimal 2 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@353 -- # local d=2 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@355 -- # echo 2 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@366 -- # ver2[v]=2 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@368 -- # return 0 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:27:02.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:02.805 --rc genhtml_branch_coverage=1 00:27:02.805 --rc genhtml_function_coverage=1 00:27:02.805 --rc genhtml_legend=1 00:27:02.805 --rc geninfo_all_blocks=1 00:27:02.805 --rc geninfo_unexecuted_blocks=1 00:27:02.805 00:27:02.805 ' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:27:02.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:02.805 --rc genhtml_branch_coverage=1 00:27:02.805 --rc genhtml_function_coverage=1 00:27:02.805 --rc genhtml_legend=1 00:27:02.805 --rc geninfo_all_blocks=1 00:27:02.805 --rc geninfo_unexecuted_blocks=1 00:27:02.805 00:27:02.805 ' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:27:02.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:02.805 --rc genhtml_branch_coverage=1 00:27:02.805 --rc genhtml_function_coverage=1 00:27:02.805 --rc genhtml_legend=1 00:27:02.805 --rc geninfo_all_blocks=1 00:27:02.805 --rc geninfo_unexecuted_blocks=1 00:27:02.805 00:27:02.805 ' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:27:02.805 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:02.805 --rc genhtml_branch_coverage=1 00:27:02.805 --rc genhtml_function_coverage=1 00:27:02.805 --rc genhtml_legend=1 00:27:02.805 --rc geninfo_all_blocks=1 00:27:02.805 --rc geninfo_unexecuted_blocks=1 00:27:02.805 00:27:02.805 ' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@7 -- # uname -s 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@15 -- # shopt -s extglob 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@5 -- # export PATH 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@51 -- # : 0 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:02.805 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@11 -- # nvmftestinit 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@474 -- # prepare_net_devs 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@436 -- # local -g is_hw=no 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@438 -- # remove_spdk_ns 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@309 -- # xtrace_disable 00:27:02.805 09:02:52 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@315 -- # pci_devs=() 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@315 -- # local -a pci_devs 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@316 -- # pci_net_devs=() 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@317 -- # pci_drivers=() 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@317 -- # local -A pci_drivers 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@319 -- # net_devs=() 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@319 -- # local -ga net_devs 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@320 -- # e810=() 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@320 -- # local -ga e810 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@321 -- # x722=() 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@321 -- # local -ga x722 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@322 -- # mlx=() 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@322 -- # local -ga mlx 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:27:10.943 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:27:10.943 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@416 -- # [[ up == up ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:27:10.943 Found net devices under 0000:4b:00.0: cvl_0_0 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@416 -- # [[ up == up ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:27:10.943 Found net devices under 0000:4b:00.1: cvl_0_1 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@440 -- # is_hw=yes 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:27:10.943 09:02:59 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:10.943 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:10.943 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:10.943 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:27:10.943 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:27:10.943 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:10.943 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.566 ms 00:27:10.943 00:27:10.943 --- 10.0.0.2 ping statistics --- 00:27:10.943 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:10.943 rtt min/avg/max/mdev = 0.566/0.566/0.566/0.000 ms 00:27:10.943 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:10.943 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:10.943 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.285 ms 00:27:10.943 00:27:10.943 --- 10.0.0.1 ping statistics --- 00:27:10.943 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:10.943 rtt min/avg/max/mdev = 0.285/0.285/0.285/0.000 ms 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@448 -- # return 0 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@12 -- # nvmfappstart -m 0xF 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@724 -- # xtrace_disable 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@507 -- # nvmfpid=1975601 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@508 -- # waitforlisten 1975601 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@831 -- # '[' -z 1975601 ']' 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:10.944 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:10.944 [2024-11-06 09:03:00.191832] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:27:10.944 [2024-11-06 09:03:00.191890] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:10.944 [2024-11-06 09:03:00.270770] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:27:10.944 [2024-11-06 09:03:00.307782] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:10.944 [2024-11-06 09:03:00.307818] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:10.944 [2024-11-06 09:03:00.307826] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:10.944 [2024-11-06 09:03:00.307832] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:10.944 [2024-11-06 09:03:00.307838] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:10.944 [2024-11-06 09:03:00.309491] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:10.944 [2024-11-06 09:03:00.309604] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:27:10.944 [2024-11-06 09:03:00.309774] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:27:10.944 [2024-11-06 09:03:00.309797] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@864 -- # return 0 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@730 -- # xtrace_disable 00:27:10.944 09:03:00 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:10.944 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:10.944 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@14 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:27:10.944 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.944 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:10.944 [2024-11-06 09:03:01.032983] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:10.944 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:10.944 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@16 -- # rpc_cmd bdev_malloc_create 64 512 --name Malloc0 00:27:10.944 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:10.944 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.205 Malloc0 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@17 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 2 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@18 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@19 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.205 [2024-11-06 09:03:01.107146] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@21 -- # rpc_cmd nvmf_get_subsystems 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.205 [ 00:27:11.205 { 00:27:11.205 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:27:11.205 "subtype": "Discovery", 00:27:11.205 "listen_addresses": [], 00:27:11.205 "allow_any_host": true, 00:27:11.205 "hosts": [] 00:27:11.205 }, 00:27:11.205 { 00:27:11.205 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:11.205 "subtype": "NVMe", 00:27:11.205 "listen_addresses": [ 00:27:11.205 { 00:27:11.205 "trtype": "TCP", 00:27:11.205 "adrfam": "IPv4", 00:27:11.205 "traddr": "10.0.0.2", 00:27:11.205 "trsvcid": "4420" 00:27:11.205 } 00:27:11.205 ], 00:27:11.205 "allow_any_host": true, 00:27:11.205 "hosts": [], 00:27:11.205 "serial_number": "SPDK00000000000001", 00:27:11.205 "model_number": "SPDK bdev Controller", 00:27:11.205 "max_namespaces": 2, 00:27:11.205 "min_cntlid": 1, 00:27:11.205 "max_cntlid": 65519, 00:27:11.205 "namespaces": [ 00:27:11.205 { 00:27:11.205 "nsid": 1, 00:27:11.205 "bdev_name": "Malloc0", 00:27:11.205 "name": "Malloc0", 00:27:11.205 "nguid": "C754F44C46DF4EC19F6BAE5EEB1FAC68", 00:27:11.205 "uuid": "c754f44c-46df-4ec1-9f6b-ae5eeb1fac68" 00:27:11.205 } 00:27:11.205 ] 00:27:11.205 } 00:27:11.205 ] 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@23 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@24 -- # rm -f /tmp/aer_touch_file 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@33 -- # aerpid=1975900 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -n 2 -t /tmp/aer_touch_file 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@36 -- # waitforfile /tmp/aer_touch_file 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1265 -- # local i=0 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1267 -- # '[' 0 -lt 200 ']' 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1268 -- # i=1 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1269 -- # sleep 0.1 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1267 -- # '[' 1 -lt 200 ']' 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1268 -- # i=2 00:27:11.205 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1269 -- # sleep 0.1 00:27:11.465 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:27:11.465 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1267 -- # '[' 2 -lt 200 ']' 00:27:11.465 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1268 -- # i=3 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1269 -- # sleep 0.1 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1272 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1276 -- # return 0 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@39 -- # rpc_cmd bdev_malloc_create 64 4096 --name Malloc1 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.466 Malloc1 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@40 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 2 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@41 -- # rpc_cmd nvmf_get_subsystems 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.466 Asynchronous Event Request test 00:27:11.466 Attaching to 10.0.0.2 00:27:11.466 Attached to 10.0.0.2 00:27:11.466 Registering asynchronous event callbacks... 00:27:11.466 Starting namespace attribute notice tests for all controllers... 00:27:11.466 10.0.0.2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:27:11.466 aer_cb - Changed Namespace 00:27:11.466 Cleaning up... 00:27:11.466 [ 00:27:11.466 { 00:27:11.466 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:27:11.466 "subtype": "Discovery", 00:27:11.466 "listen_addresses": [], 00:27:11.466 "allow_any_host": true, 00:27:11.466 "hosts": [] 00:27:11.466 }, 00:27:11.466 { 00:27:11.466 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:11.466 "subtype": "NVMe", 00:27:11.466 "listen_addresses": [ 00:27:11.466 { 00:27:11.466 "trtype": "TCP", 00:27:11.466 "adrfam": "IPv4", 00:27:11.466 "traddr": "10.0.0.2", 00:27:11.466 "trsvcid": "4420" 00:27:11.466 } 00:27:11.466 ], 00:27:11.466 "allow_any_host": true, 00:27:11.466 "hosts": [], 00:27:11.466 "serial_number": "SPDK00000000000001", 00:27:11.466 "model_number": "SPDK bdev Controller", 00:27:11.466 "max_namespaces": 2, 00:27:11.466 "min_cntlid": 1, 00:27:11.466 "max_cntlid": 65519, 00:27:11.466 "namespaces": [ 00:27:11.466 { 00:27:11.466 "nsid": 1, 00:27:11.466 "bdev_name": "Malloc0", 00:27:11.466 "name": "Malloc0", 00:27:11.466 "nguid": "C754F44C46DF4EC19F6BAE5EEB1FAC68", 00:27:11.466 "uuid": "c754f44c-46df-4ec1-9f6b-ae5eeb1fac68" 00:27:11.466 }, 00:27:11.466 { 00:27:11.466 "nsid": 2, 00:27:11.466 "bdev_name": "Malloc1", 00:27:11.466 "name": "Malloc1", 00:27:11.466 "nguid": "4520668202774E25AE1F987775521CE0", 00:27:11.466 "uuid": "45206682-0277-4e25-ae1f-987775521ce0" 00:27:11.466 } 00:27:11.466 ] 00:27:11.466 } 00:27:11.466 ] 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@43 -- # wait 1975900 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@45 -- # rpc_cmd bdev_malloc_delete Malloc0 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@46 -- # rpc_cmd bdev_malloc_delete Malloc1 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@47 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@49 -- # trap - SIGINT SIGTERM EXIT 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- host/aer.sh@51 -- # nvmftestfini 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@514 -- # nvmfcleanup 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@121 -- # sync 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@124 -- # set +e 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:11.466 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:11.466 rmmod nvme_tcp 00:27:11.726 rmmod nvme_fabrics 00:27:11.726 rmmod nvme_keyring 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@128 -- # set -e 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@129 -- # return 0 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@515 -- # '[' -n 1975601 ']' 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@516 -- # killprocess 1975601 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@950 -- # '[' -z 1975601 ']' 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@954 -- # kill -0 1975601 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@955 -- # uname 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1975601 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1975601' 00:27:11.727 killing process with pid 1975601 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@969 -- # kill 1975601 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@974 -- # wait 1975601 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@297 -- # iptr 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@789 -- # iptables-save 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@789 -- # iptables-restore 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:11.727 09:03:01 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:14.269 09:03:03 nvmf_tcp.nvmf_host.nvmf_aer -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:14.269 00:27:14.269 real 0m11.453s 00:27:14.269 user 0m8.304s 00:27:14.269 sys 0m6.026s 00:27:14.269 09:03:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@1126 -- # xtrace_disable 00:27:14.269 09:03:03 nvmf_tcp.nvmf_host.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:27:14.269 ************************************ 00:27:14.269 END TEST nvmf_aer 00:27:14.269 ************************************ 00:27:14.269 09:03:03 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@18 -- # run_test nvmf_async_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:27:14.269 09:03:03 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:27:14.269 09:03:03 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:27:14.269 09:03:03 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:27:14.269 ************************************ 00:27:14.269 START TEST nvmf_async_init 00:27:14.269 ************************************ 00:27:14.269 09:03:03 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:27:14.269 * Looking for test storage... 00:27:14.269 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:27:14.269 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:27:14.269 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1689 -- # lcov --version 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@336 -- # IFS=.-: 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@336 -- # read -ra ver1 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@337 -- # IFS=.-: 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@337 -- # read -ra ver2 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@338 -- # local 'op=<' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@340 -- # ver1_l=2 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@341 -- # ver2_l=1 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@344 -- # case "$op" in 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@345 -- # : 1 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@365 -- # decimal 1 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@353 -- # local d=1 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@355 -- # echo 1 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@365 -- # ver1[v]=1 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@366 -- # decimal 2 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@353 -- # local d=2 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@355 -- # echo 2 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@366 -- # ver2[v]=2 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@368 -- # return 0 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:27:14.270 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:14.270 --rc genhtml_branch_coverage=1 00:27:14.270 --rc genhtml_function_coverage=1 00:27:14.270 --rc genhtml_legend=1 00:27:14.270 --rc geninfo_all_blocks=1 00:27:14.270 --rc geninfo_unexecuted_blocks=1 00:27:14.270 00:27:14.270 ' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:27:14.270 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:14.270 --rc genhtml_branch_coverage=1 00:27:14.270 --rc genhtml_function_coverage=1 00:27:14.270 --rc genhtml_legend=1 00:27:14.270 --rc geninfo_all_blocks=1 00:27:14.270 --rc geninfo_unexecuted_blocks=1 00:27:14.270 00:27:14.270 ' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:27:14.270 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:14.270 --rc genhtml_branch_coverage=1 00:27:14.270 --rc genhtml_function_coverage=1 00:27:14.270 --rc genhtml_legend=1 00:27:14.270 --rc geninfo_all_blocks=1 00:27:14.270 --rc geninfo_unexecuted_blocks=1 00:27:14.270 00:27:14.270 ' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:27:14.270 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:14.270 --rc genhtml_branch_coverage=1 00:27:14.270 --rc genhtml_function_coverage=1 00:27:14.270 --rc genhtml_legend=1 00:27:14.270 --rc geninfo_all_blocks=1 00:27:14.270 --rc geninfo_unexecuted_blocks=1 00:27:14.270 00:27:14.270 ' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@7 -- # uname -s 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@15 -- # shopt -s extglob 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@5 -- # export PATH 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@51 -- # : 0 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:14.270 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@13 -- # null_bdev_size=1024 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@14 -- # null_block_size=512 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@15 -- # null_bdev=null0 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@16 -- # nvme_bdev=nvme0 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # uuidgen 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # tr -d - 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@20 -- # nguid=bdcd4e4cbdd549c596c0610a4180e39f 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@22 -- # nvmftestinit 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:27:14.270 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:14.271 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@474 -- # prepare_net_devs 00:27:14.271 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@436 -- # local -g is_hw=no 00:27:14.271 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@438 -- # remove_spdk_ns 00:27:14.271 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:14.271 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:14.271 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:14.271 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:27:14.271 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:27:14.271 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@309 -- # xtrace_disable 00:27:14.271 09:03:04 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@315 -- # pci_devs=() 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@315 -- # local -a pci_devs 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@316 -- # pci_net_devs=() 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@317 -- # pci_drivers=() 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@317 -- # local -A pci_drivers 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@319 -- # net_devs=() 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@319 -- # local -ga net_devs 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@320 -- # e810=() 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@320 -- # local -ga e810 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@321 -- # x722=() 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@321 -- # local -ga x722 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@322 -- # mlx=() 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@322 -- # local -ga mlx 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:27:22.405 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:27:22.405 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@416 -- # [[ up == up ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:27:22.405 Found net devices under 0000:4b:00.0: cvl_0_0 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@416 -- # [[ up == up ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:27:22.405 Found net devices under 0000:4b:00.1: cvl_0_1 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@440 -- # is_hw=yes 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:27:22.405 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:22.405 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.602 ms 00:27:22.405 00:27:22.405 --- 10.0.0.2 ping statistics --- 00:27:22.405 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:22.405 rtt min/avg/max/mdev = 0.602/0.602/0.602/0.000 ms 00:27:22.405 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:22.405 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:22.405 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.213 ms 00:27:22.405 00:27:22.405 --- 10.0.0.1 ping statistics --- 00:27:22.406 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:22.406 rtt min/avg/max/mdev = 0.213/0.213/0.213/0.000 ms 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@448 -- # return 0 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@23 -- # nvmfappstart -m 0x1 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@724 -- # xtrace_disable 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@507 -- # nvmfpid=1980678 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@508 -- # waitforlisten 1980678 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@831 -- # '[' -z 1980678 ']' 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:22.406 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:22.406 09:03:11 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.406 [2024-11-06 09:03:11.791917] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:27:22.406 [2024-11-06 09:03:11.791981] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:22.406 [2024-11-06 09:03:11.876734] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:22.406 [2024-11-06 09:03:11.920201] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:22.406 [2024-11-06 09:03:11.920244] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:22.406 [2024-11-06 09:03:11.920252] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:22.406 [2024-11-06 09:03:11.920259] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:22.406 [2024-11-06 09:03:11.920265] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:22.406 [2024-11-06 09:03:11.920900] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@864 -- # return 0 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@730 -- # xtrace_disable 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@26 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.666 [2024-11-06 09:03:12.649890] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@27 -- # rpc_cmd bdev_null_create null0 1024 512 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.666 null0 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@28 -- # rpc_cmd bdev_wait_for_examine 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@29 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 -g bdcd4e4cbdd549c596c0610a4180e39f 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@31 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.666 [2024-11-06 09:03:12.710189] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@37 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode0 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.666 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.926 nvme0n1 00:27:22.926 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.926 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@41 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:27:22.926 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.926 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.926 [ 00:27:22.926 { 00:27:22.926 "name": "nvme0n1", 00:27:22.926 "aliases": [ 00:27:22.926 "bdcd4e4c-bdd5-49c5-96c0-610a4180e39f" 00:27:22.926 ], 00:27:22.926 "product_name": "NVMe disk", 00:27:22.926 "block_size": 512, 00:27:22.926 "num_blocks": 2097152, 00:27:22.926 "uuid": "bdcd4e4c-bdd5-49c5-96c0-610a4180e39f", 00:27:22.926 "numa_id": 0, 00:27:22.926 "assigned_rate_limits": { 00:27:22.926 "rw_ios_per_sec": 0, 00:27:22.926 "rw_mbytes_per_sec": 0, 00:27:22.926 "r_mbytes_per_sec": 0, 00:27:22.926 "w_mbytes_per_sec": 0 00:27:22.926 }, 00:27:22.926 "claimed": false, 00:27:22.926 "zoned": false, 00:27:22.926 "supported_io_types": { 00:27:22.926 "read": true, 00:27:22.926 "write": true, 00:27:22.926 "unmap": false, 00:27:22.926 "flush": true, 00:27:22.926 "reset": true, 00:27:22.926 "nvme_admin": true, 00:27:22.926 "nvme_io": true, 00:27:22.926 "nvme_io_md": false, 00:27:22.926 "write_zeroes": true, 00:27:22.926 "zcopy": false, 00:27:22.926 "get_zone_info": false, 00:27:22.926 "zone_management": false, 00:27:22.926 "zone_append": false, 00:27:22.926 "compare": true, 00:27:22.926 "compare_and_write": true, 00:27:22.926 "abort": true, 00:27:22.926 "seek_hole": false, 00:27:22.926 "seek_data": false, 00:27:22.926 "copy": true, 00:27:22.926 "nvme_iov_md": false 00:27:22.926 }, 00:27:22.926 "memory_domains": [ 00:27:22.926 { 00:27:22.926 "dma_device_id": "system", 00:27:22.926 "dma_device_type": 1 00:27:22.926 } 00:27:22.926 ], 00:27:22.926 "driver_specific": { 00:27:22.926 "nvme": [ 00:27:22.926 { 00:27:22.926 "trid": { 00:27:22.926 "trtype": "TCP", 00:27:22.926 "adrfam": "IPv4", 00:27:22.926 "traddr": "10.0.0.2", 00:27:22.926 "trsvcid": "4420", 00:27:22.926 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:27:22.926 }, 00:27:22.926 "ctrlr_data": { 00:27:22.926 "cntlid": 1, 00:27:22.926 "vendor_id": "0x8086", 00:27:22.926 "model_number": "SPDK bdev Controller", 00:27:22.926 "serial_number": "00000000000000000000", 00:27:22.926 "firmware_revision": "25.01", 00:27:22.926 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:27:22.926 "oacs": { 00:27:22.926 "security": 0, 00:27:22.926 "format": 0, 00:27:22.926 "firmware": 0, 00:27:22.926 "ns_manage": 0 00:27:22.926 }, 00:27:22.926 "multi_ctrlr": true, 00:27:22.926 "ana_reporting": false 00:27:22.926 }, 00:27:22.926 "vs": { 00:27:22.926 "nvme_version": "1.3" 00:27:22.926 }, 00:27:22.926 "ns_data": { 00:27:22.926 "id": 1, 00:27:22.926 "can_share": true 00:27:22.926 } 00:27:22.926 } 00:27:22.926 ], 00:27:22.926 "mp_policy": "active_passive" 00:27:22.926 } 00:27:22.926 } 00:27:22.926 ] 00:27:22.926 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:22.926 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@44 -- # rpc_cmd bdev_nvme_reset_controller nvme0 00:27:22.927 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:22.927 09:03:12 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:22.927 [2024-11-06 09:03:12.987402] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:27:22.927 [2024-11-06 09:03:12.987469] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1261ce0 (9): Bad file descriptor 00:27:23.187 [2024-11-06 09:03:13.119852] bdev_nvme.c:2236:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 2] Resetting controller successful. 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@47 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:23.187 [ 00:27:23.187 { 00:27:23.187 "name": "nvme0n1", 00:27:23.187 "aliases": [ 00:27:23.187 "bdcd4e4c-bdd5-49c5-96c0-610a4180e39f" 00:27:23.187 ], 00:27:23.187 "product_name": "NVMe disk", 00:27:23.187 "block_size": 512, 00:27:23.187 "num_blocks": 2097152, 00:27:23.187 "uuid": "bdcd4e4c-bdd5-49c5-96c0-610a4180e39f", 00:27:23.187 "numa_id": 0, 00:27:23.187 "assigned_rate_limits": { 00:27:23.187 "rw_ios_per_sec": 0, 00:27:23.187 "rw_mbytes_per_sec": 0, 00:27:23.187 "r_mbytes_per_sec": 0, 00:27:23.187 "w_mbytes_per_sec": 0 00:27:23.187 }, 00:27:23.187 "claimed": false, 00:27:23.187 "zoned": false, 00:27:23.187 "supported_io_types": { 00:27:23.187 "read": true, 00:27:23.187 "write": true, 00:27:23.187 "unmap": false, 00:27:23.187 "flush": true, 00:27:23.187 "reset": true, 00:27:23.187 "nvme_admin": true, 00:27:23.187 "nvme_io": true, 00:27:23.187 "nvme_io_md": false, 00:27:23.187 "write_zeroes": true, 00:27:23.187 "zcopy": false, 00:27:23.187 "get_zone_info": false, 00:27:23.187 "zone_management": false, 00:27:23.187 "zone_append": false, 00:27:23.187 "compare": true, 00:27:23.187 "compare_and_write": true, 00:27:23.187 "abort": true, 00:27:23.187 "seek_hole": false, 00:27:23.187 "seek_data": false, 00:27:23.187 "copy": true, 00:27:23.187 "nvme_iov_md": false 00:27:23.187 }, 00:27:23.187 "memory_domains": [ 00:27:23.187 { 00:27:23.187 "dma_device_id": "system", 00:27:23.187 "dma_device_type": 1 00:27:23.187 } 00:27:23.187 ], 00:27:23.187 "driver_specific": { 00:27:23.187 "nvme": [ 00:27:23.187 { 00:27:23.187 "trid": { 00:27:23.187 "trtype": "TCP", 00:27:23.187 "adrfam": "IPv4", 00:27:23.187 "traddr": "10.0.0.2", 00:27:23.187 "trsvcid": "4420", 00:27:23.187 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:27:23.187 }, 00:27:23.187 "ctrlr_data": { 00:27:23.187 "cntlid": 2, 00:27:23.187 "vendor_id": "0x8086", 00:27:23.187 "model_number": "SPDK bdev Controller", 00:27:23.187 "serial_number": "00000000000000000000", 00:27:23.187 "firmware_revision": "25.01", 00:27:23.187 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:27:23.187 "oacs": { 00:27:23.187 "security": 0, 00:27:23.187 "format": 0, 00:27:23.187 "firmware": 0, 00:27:23.187 "ns_manage": 0 00:27:23.187 }, 00:27:23.187 "multi_ctrlr": true, 00:27:23.187 "ana_reporting": false 00:27:23.187 }, 00:27:23.187 "vs": { 00:27:23.187 "nvme_version": "1.3" 00:27:23.187 }, 00:27:23.187 "ns_data": { 00:27:23.187 "id": 1, 00:27:23.187 "can_share": true 00:27:23.187 } 00:27:23.187 } 00:27:23.187 ], 00:27:23.187 "mp_policy": "active_passive" 00:27:23.187 } 00:27:23.187 } 00:27:23.187 ] 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@50 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@53 -- # mktemp 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@53 -- # key_path=/tmp/tmp.vSSWvLl5ca 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@54 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@55 -- # chmod 0600 /tmp/tmp.vSSWvLl5ca 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@56 -- # rpc_cmd keyring_file_add_key key0 /tmp/tmp.vSSWvLl5ca 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@57 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode0 --disable 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@58 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 --secure-channel 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:23.187 [2024-11-06 09:03:13.212104] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:27:23.187 [2024-11-06 09:03:13.212217] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@60 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host1 --psk key0 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@66 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4421 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host1 --psk key0 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.187 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:23.187 [2024-11-06 09:03:13.236187] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:27:23.448 nvme0n1 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@70 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:23.448 [ 00:27:23.448 { 00:27:23.448 "name": "nvme0n1", 00:27:23.448 "aliases": [ 00:27:23.448 "bdcd4e4c-bdd5-49c5-96c0-610a4180e39f" 00:27:23.448 ], 00:27:23.448 "product_name": "NVMe disk", 00:27:23.448 "block_size": 512, 00:27:23.448 "num_blocks": 2097152, 00:27:23.448 "uuid": "bdcd4e4c-bdd5-49c5-96c0-610a4180e39f", 00:27:23.448 "numa_id": 0, 00:27:23.448 "assigned_rate_limits": { 00:27:23.448 "rw_ios_per_sec": 0, 00:27:23.448 "rw_mbytes_per_sec": 0, 00:27:23.448 "r_mbytes_per_sec": 0, 00:27:23.448 "w_mbytes_per_sec": 0 00:27:23.448 }, 00:27:23.448 "claimed": false, 00:27:23.448 "zoned": false, 00:27:23.448 "supported_io_types": { 00:27:23.448 "read": true, 00:27:23.448 "write": true, 00:27:23.448 "unmap": false, 00:27:23.448 "flush": true, 00:27:23.448 "reset": true, 00:27:23.448 "nvme_admin": true, 00:27:23.448 "nvme_io": true, 00:27:23.448 "nvme_io_md": false, 00:27:23.448 "write_zeroes": true, 00:27:23.448 "zcopy": false, 00:27:23.448 "get_zone_info": false, 00:27:23.448 "zone_management": false, 00:27:23.448 "zone_append": false, 00:27:23.448 "compare": true, 00:27:23.448 "compare_and_write": true, 00:27:23.448 "abort": true, 00:27:23.448 "seek_hole": false, 00:27:23.448 "seek_data": false, 00:27:23.448 "copy": true, 00:27:23.448 "nvme_iov_md": false 00:27:23.448 }, 00:27:23.448 "memory_domains": [ 00:27:23.448 { 00:27:23.448 "dma_device_id": "system", 00:27:23.448 "dma_device_type": 1 00:27:23.448 } 00:27:23.448 ], 00:27:23.448 "driver_specific": { 00:27:23.448 "nvme": [ 00:27:23.448 { 00:27:23.448 "trid": { 00:27:23.448 "trtype": "TCP", 00:27:23.448 "adrfam": "IPv4", 00:27:23.448 "traddr": "10.0.0.2", 00:27:23.448 "trsvcid": "4421", 00:27:23.448 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:27:23.448 }, 00:27:23.448 "ctrlr_data": { 00:27:23.448 "cntlid": 3, 00:27:23.448 "vendor_id": "0x8086", 00:27:23.448 "model_number": "SPDK bdev Controller", 00:27:23.448 "serial_number": "00000000000000000000", 00:27:23.448 "firmware_revision": "25.01", 00:27:23.448 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:27:23.448 "oacs": { 00:27:23.448 "security": 0, 00:27:23.448 "format": 0, 00:27:23.448 "firmware": 0, 00:27:23.448 "ns_manage": 0 00:27:23.448 }, 00:27:23.448 "multi_ctrlr": true, 00:27:23.448 "ana_reporting": false 00:27:23.448 }, 00:27:23.448 "vs": { 00:27:23.448 "nvme_version": "1.3" 00:27:23.448 }, 00:27:23.448 "ns_data": { 00:27:23.448 "id": 1, 00:27:23.448 "can_share": true 00:27:23.448 } 00:27:23.448 } 00:27:23.448 ], 00:27:23.448 "mp_policy": "active_passive" 00:27:23.448 } 00:27:23.448 } 00:27:23.448 ] 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@73 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@76 -- # rm -f /tmp/tmp.vSSWvLl5ca 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@78 -- # trap - SIGINT SIGTERM EXIT 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- host/async_init.sh@79 -- # nvmftestfini 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@514 -- # nvmfcleanup 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@121 -- # sync 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@124 -- # set +e 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:23.448 rmmod nvme_tcp 00:27:23.448 rmmod nvme_fabrics 00:27:23.448 rmmod nvme_keyring 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@128 -- # set -e 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@129 -- # return 0 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@515 -- # '[' -n 1980678 ']' 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@516 -- # killprocess 1980678 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@950 -- # '[' -z 1980678 ']' 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@954 -- # kill -0 1980678 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@955 -- # uname 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1980678 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1980678' 00:27:23.448 killing process with pid 1980678 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@969 -- # kill 1980678 00:27:23.448 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@974 -- # wait 1980678 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@297 -- # iptr 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@789 -- # iptables-save 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@789 -- # iptables-restore 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:23.709 09:03:13 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:25.625 09:03:15 nvmf_tcp.nvmf_host.nvmf_async_init -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:25.625 00:27:25.625 real 0m11.722s 00:27:25.625 user 0m4.215s 00:27:25.625 sys 0m6.046s 00:27:25.625 09:03:15 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@1126 -- # xtrace_disable 00:27:25.625 09:03:15 nvmf_tcp.nvmf_host.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:27:25.625 ************************************ 00:27:25.625 END TEST nvmf_async_init 00:27:25.625 ************************************ 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@19 -- # run_test dma /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:27:25.886 ************************************ 00:27:25.886 START TEST dma 00:27:25.886 ************************************ 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:27:25.886 * Looking for test storage... 00:27:25.886 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1689 -- # lcov --version 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@336 -- # IFS=.-: 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@336 -- # read -ra ver1 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@337 -- # IFS=.-: 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@337 -- # read -ra ver2 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@338 -- # local 'op=<' 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@340 -- # ver1_l=2 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@341 -- # ver2_l=1 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@344 -- # case "$op" in 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@345 -- # : 1 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@365 -- # decimal 1 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@353 -- # local d=1 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@355 -- # echo 1 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@365 -- # ver1[v]=1 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@366 -- # decimal 2 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@353 -- # local d=2 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:25.886 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@355 -- # echo 2 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@366 -- # ver2[v]=2 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@368 -- # return 0 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:27:25.887 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:25.887 --rc genhtml_branch_coverage=1 00:27:25.887 --rc genhtml_function_coverage=1 00:27:25.887 --rc genhtml_legend=1 00:27:25.887 --rc geninfo_all_blocks=1 00:27:25.887 --rc geninfo_unexecuted_blocks=1 00:27:25.887 00:27:25.887 ' 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:27:25.887 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:25.887 --rc genhtml_branch_coverage=1 00:27:25.887 --rc genhtml_function_coverage=1 00:27:25.887 --rc genhtml_legend=1 00:27:25.887 --rc geninfo_all_blocks=1 00:27:25.887 --rc geninfo_unexecuted_blocks=1 00:27:25.887 00:27:25.887 ' 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:27:25.887 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:25.887 --rc genhtml_branch_coverage=1 00:27:25.887 --rc genhtml_function_coverage=1 00:27:25.887 --rc genhtml_legend=1 00:27:25.887 --rc geninfo_all_blocks=1 00:27:25.887 --rc geninfo_unexecuted_blocks=1 00:27:25.887 00:27:25.887 ' 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:27:25.887 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:25.887 --rc genhtml_branch_coverage=1 00:27:25.887 --rc genhtml_function_coverage=1 00:27:25.887 --rc genhtml_legend=1 00:27:25.887 --rc geninfo_all_blocks=1 00:27:25.887 --rc geninfo_unexecuted_blocks=1 00:27:25.887 00:27:25.887 ' 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- host/dma.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@7 -- # uname -s 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@15 -- # shopt -s extglob 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:25.887 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- paths/export.sh@5 -- # export PATH 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@51 -- # : 0 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:26.149 09:03:15 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:26.149 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host.dma -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host.dma -- host/dma.sh@12 -- # '[' tcp '!=' rdma ']' 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host.dma -- host/dma.sh@13 -- # exit 0 00:27:26.149 00:27:26.149 real 0m0.229s 00:27:26.149 user 0m0.131s 00:27:26.149 sys 0m0.112s 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@1126 -- # xtrace_disable 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host.dma -- common/autotest_common.sh@10 -- # set +x 00:27:26.149 ************************************ 00:27:26.149 END TEST dma 00:27:26.149 ************************************ 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@22 -- # run_test nvmf_identify /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:27:26.149 09:03:16 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:27:26.149 ************************************ 00:27:26.149 START TEST nvmf_identify 00:27:26.149 ************************************ 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:27:26.150 * Looking for test storage... 00:27:26.150 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1689 -- # lcov --version 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@336 -- # IFS=.-: 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@336 -- # read -ra ver1 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@337 -- # IFS=.-: 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@337 -- # read -ra ver2 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@338 -- # local 'op=<' 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@340 -- # ver1_l=2 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@341 -- # ver2_l=1 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@344 -- # case "$op" in 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@345 -- # : 1 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@365 -- # decimal 1 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@353 -- # local d=1 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@355 -- # echo 1 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@365 -- # ver1[v]=1 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@366 -- # decimal 2 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@353 -- # local d=2 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@355 -- # echo 2 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@366 -- # ver2[v]=2 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@368 -- # return 0 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:27:26.150 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:26.150 --rc genhtml_branch_coverage=1 00:27:26.150 --rc genhtml_function_coverage=1 00:27:26.150 --rc genhtml_legend=1 00:27:26.150 --rc geninfo_all_blocks=1 00:27:26.150 --rc geninfo_unexecuted_blocks=1 00:27:26.150 00:27:26.150 ' 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:27:26.150 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:26.150 --rc genhtml_branch_coverage=1 00:27:26.150 --rc genhtml_function_coverage=1 00:27:26.150 --rc genhtml_legend=1 00:27:26.150 --rc geninfo_all_blocks=1 00:27:26.150 --rc geninfo_unexecuted_blocks=1 00:27:26.150 00:27:26.150 ' 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:27:26.150 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:26.150 --rc genhtml_branch_coverage=1 00:27:26.150 --rc genhtml_function_coverage=1 00:27:26.150 --rc genhtml_legend=1 00:27:26.150 --rc geninfo_all_blocks=1 00:27:26.150 --rc geninfo_unexecuted_blocks=1 00:27:26.150 00:27:26.150 ' 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:27:26.150 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:26.150 --rc genhtml_branch_coverage=1 00:27:26.150 --rc genhtml_function_coverage=1 00:27:26.150 --rc genhtml_legend=1 00:27:26.150 --rc geninfo_all_blocks=1 00:27:26.150 --rc geninfo_unexecuted_blocks=1 00:27:26.150 00:27:26.150 ' 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:26.150 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@7 -- # uname -s 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@15 -- # shopt -s extglob 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@5 -- # export PATH 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@51 -- # : 0 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:26.412 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@11 -- # MALLOC_BDEV_SIZE=64 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@14 -- # nvmftestinit 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@474 -- # prepare_net_devs 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@436 -- # local -g is_hw=no 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@438 -- # remove_spdk_ns 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@309 -- # xtrace_disable 00:27:26.412 09:03:16 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@315 -- # pci_devs=() 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@315 -- # local -a pci_devs 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@316 -- # pci_net_devs=() 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@317 -- # pci_drivers=() 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@317 -- # local -A pci_drivers 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@319 -- # net_devs=() 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@319 -- # local -ga net_devs 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@320 -- # e810=() 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@320 -- # local -ga e810 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@321 -- # x722=() 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@321 -- # local -ga x722 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@322 -- # mlx=() 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@322 -- # local -ga mlx 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:27:32.999 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:27:33.000 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:27:33.000 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@416 -- # [[ up == up ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:27:33.000 Found net devices under 0000:4b:00.0: cvl_0_0 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@416 -- # [[ up == up ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:27:33.000 Found net devices under 0000:4b:00.1: cvl_0_1 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@440 -- # is_hw=yes 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:27:33.000 09:03:22 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:27:33.000 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:27:33.000 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:27:33.261 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:33.261 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.589 ms 00:27:33.261 00:27:33.261 --- 10.0.0.2 ping statistics --- 00:27:33.261 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:33.261 rtt min/avg/max/mdev = 0.589/0.589/0.589/0.000 ms 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:33.261 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:33.261 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.307 ms 00:27:33.261 00:27:33.261 --- 10.0.0.1 ping statistics --- 00:27:33.261 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:33.261 rtt min/avg/max/mdev = 0.307/0.307/0.307/0.000 ms 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@448 -- # return 0 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@16 -- # timing_enter start_nvmf_tgt 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@724 -- # xtrace_disable 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@19 -- # nvmfpid=1985250 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@21 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@18 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@23 -- # waitforlisten 1985250 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@831 -- # '[' -z 1985250 ']' 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:33.261 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:33.261 09:03:23 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:33.261 [2024-11-06 09:03:23.370254] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:27:33.261 [2024-11-06 09:03:23.370303] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:33.523 [2024-11-06 09:03:23.451203] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:27:33.523 [2024-11-06 09:03:23.487668] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:33.523 [2024-11-06 09:03:23.487702] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:33.523 [2024-11-06 09:03:23.487710] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:33.523 [2024-11-06 09:03:23.487717] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:33.523 [2024-11-06 09:03:23.487726] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:33.523 [2024-11-06 09:03:23.489248] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:33.523 [2024-11-06 09:03:23.489264] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:27:33.523 [2024-11-06 09:03:23.489397] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:33.523 [2024-11-06 09:03:23.489398] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:27:34.094 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:34.094 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@864 -- # return 0 00:27:34.094 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@24 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:27:34.094 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.094 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:34.094 [2024-11-06 09:03:24.176395] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:34.094 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.094 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@25 -- # timing_exit start_nvmf_tgt 00:27:34.094 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@730 -- # xtrace_disable 00:27:34.094 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@27 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:34.356 Malloc0 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 --nguid ABCDEF0123456789ABCDEF0123456789 --eui64 ABCDEF0123456789 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:34.356 [2024-11-06 09:03:24.291154] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@35 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@37 -- # rpc_cmd nvmf_get_subsystems 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:34.356 [ 00:27:34.356 { 00:27:34.356 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:27:34.356 "subtype": "Discovery", 00:27:34.356 "listen_addresses": [ 00:27:34.356 { 00:27:34.356 "trtype": "TCP", 00:27:34.356 "adrfam": "IPv4", 00:27:34.356 "traddr": "10.0.0.2", 00:27:34.356 "trsvcid": "4420" 00:27:34.356 } 00:27:34.356 ], 00:27:34.356 "allow_any_host": true, 00:27:34.356 "hosts": [] 00:27:34.356 }, 00:27:34.356 { 00:27:34.356 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:34.356 "subtype": "NVMe", 00:27:34.356 "listen_addresses": [ 00:27:34.356 { 00:27:34.356 "trtype": "TCP", 00:27:34.356 "adrfam": "IPv4", 00:27:34.356 "traddr": "10.0.0.2", 00:27:34.356 "trsvcid": "4420" 00:27:34.356 } 00:27:34.356 ], 00:27:34.356 "allow_any_host": true, 00:27:34.356 "hosts": [], 00:27:34.356 "serial_number": "SPDK00000000000001", 00:27:34.356 "model_number": "SPDK bdev Controller", 00:27:34.356 "max_namespaces": 32, 00:27:34.356 "min_cntlid": 1, 00:27:34.356 "max_cntlid": 65519, 00:27:34.356 "namespaces": [ 00:27:34.356 { 00:27:34.356 "nsid": 1, 00:27:34.356 "bdev_name": "Malloc0", 00:27:34.356 "name": "Malloc0", 00:27:34.356 "nguid": "ABCDEF0123456789ABCDEF0123456789", 00:27:34.356 "eui64": "ABCDEF0123456789", 00:27:34.356 "uuid": "b44a0de2-7044-48ea-bf39-d1dbd43e4bc9" 00:27:34.356 } 00:27:34.356 ] 00:27:34.356 } 00:27:34.356 ] 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.356 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' -L all 00:27:34.356 [2024-11-06 09:03:24.354957] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:27:34.356 [2024-11-06 09:03:24.354997] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1985533 ] 00:27:34.356 [2024-11-06 09:03:24.409915] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 0] setting state to connect adminq (no timeout) 00:27:34.356 [2024-11-06 09:03:24.409973] nvme_tcp.c:2238:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:27:34.356 [2024-11-06 09:03:24.409979] nvme_tcp.c:2242:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:27:34.356 [2024-11-06 09:03:24.409994] nvme_tcp.c:2263:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:27:34.356 [2024-11-06 09:03:24.410003] sock.c: 373:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:27:34.356 [2024-11-06 09:03:24.410703] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 0] setting state to wait for connect adminq (no timeout) 00:27:34.356 [2024-11-06 09:03:24.410736] nvme_tcp.c:1455:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x1e2f690 0 00:27:34.356 [2024-11-06 09:03:24.420760] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:27:34.356 [2024-11-06 09:03:24.420773] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:27:34.356 [2024-11-06 09:03:24.420778] nvme_tcp.c:1501:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:27:34.356 [2024-11-06 09:03:24.420782] nvme_tcp.c:1502:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:27:34.356 [2024-11-06 09:03:24.420814] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.356 [2024-11-06 09:03:24.420820] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.356 [2024-11-06 09:03:24.420825] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1e2f690) 00:27:34.356 [2024-11-06 09:03:24.420838] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:27:34.356 [2024-11-06 09:03:24.420856] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91100, cid 0, qid 0 00:27:34.356 [2024-11-06 09:03:24.428757] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.356 [2024-11-06 09:03:24.428767] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.356 [2024-11-06 09:03:24.428771] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.356 [2024-11-06 09:03:24.428776] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91100) on tqpair=0x1e2f690 00:27:34.356 [2024-11-06 09:03:24.428786] nvme_fabric.c: 621:nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:27:34.356 [2024-11-06 09:03:24.428793] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read vs (no timeout) 00:27:34.356 [2024-11-06 09:03:24.428803] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read vs wait for vs (no timeout) 00:27:34.356 [2024-11-06 09:03:24.428817] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.356 [2024-11-06 09:03:24.428821] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.356 [2024-11-06 09:03:24.428825] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1e2f690) 00:27:34.356 [2024-11-06 09:03:24.428833] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.356 [2024-11-06 09:03:24.428846] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91100, cid 0, qid 0 00:27:34.356 [2024-11-06 09:03:24.429061] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.356 [2024-11-06 09:03:24.429068] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.356 [2024-11-06 09:03:24.429072] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.356 [2024-11-06 09:03:24.429076] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91100) on tqpair=0x1e2f690 00:27:34.356 [2024-11-06 09:03:24.429081] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read cap (no timeout) 00:27:34.356 [2024-11-06 09:03:24.429089] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to read cap wait for cap (no timeout) 00:27:34.356 [2024-11-06 09:03:24.429096] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.356 [2024-11-06 09:03:24.429100] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.429104] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1e2f690) 00:27:34.357 [2024-11-06 09:03:24.429110] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.357 [2024-11-06 09:03:24.429121] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91100, cid 0, qid 0 00:27:34.357 [2024-11-06 09:03:24.429294] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.357 [2024-11-06 09:03:24.429301] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.357 [2024-11-06 09:03:24.429304] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.429308] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91100) on tqpair=0x1e2f690 00:27:34.357 [2024-11-06 09:03:24.429315] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to check en (no timeout) 00:27:34.357 [2024-11-06 09:03:24.429323] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to check en wait for cc (timeout 15000 ms) 00:27:34.357 [2024-11-06 09:03:24.429329] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.429333] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.429337] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1e2f690) 00:27:34.357 [2024-11-06 09:03:24.429344] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.357 [2024-11-06 09:03:24.429354] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91100, cid 0, qid 0 00:27:34.357 [2024-11-06 09:03:24.429532] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.357 [2024-11-06 09:03:24.429539] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.357 [2024-11-06 09:03:24.429542] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.429546] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91100) on tqpair=0x1e2f690 00:27:34.357 [2024-11-06 09:03:24.429551] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:27:34.357 [2024-11-06 09:03:24.429563] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.429570] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.429574] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1e2f690) 00:27:34.357 [2024-11-06 09:03:24.429581] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.357 [2024-11-06 09:03:24.429591] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91100, cid 0, qid 0 00:27:34.357 [2024-11-06 09:03:24.429770] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.357 [2024-11-06 09:03:24.429777] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.357 [2024-11-06 09:03:24.429781] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.429785] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91100) on tqpair=0x1e2f690 00:27:34.357 [2024-11-06 09:03:24.429790] nvme_ctrlr.c:3870:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] CC.EN = 0 && CSTS.RDY = 0 00:27:34.357 [2024-11-06 09:03:24.429795] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to controller is disabled (timeout 15000 ms) 00:27:34.357 [2024-11-06 09:03:24.429802] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:27:34.357 [2024-11-06 09:03:24.429907] nvme_ctrlr.c:4068:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Setting CC.EN = 1 00:27:34.357 [2024-11-06 09:03:24.429912] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:27:34.357 [2024-11-06 09:03:24.429921] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.429925] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.429929] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1e2f690) 00:27:34.357 [2024-11-06 09:03:24.429936] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.357 [2024-11-06 09:03:24.429946] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91100, cid 0, qid 0 00:27:34.357 [2024-11-06 09:03:24.430141] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.357 [2024-11-06 09:03:24.430147] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.357 [2024-11-06 09:03:24.430151] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.430155] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91100) on tqpair=0x1e2f690 00:27:34.357 [2024-11-06 09:03:24.430160] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:27:34.357 [2024-11-06 09:03:24.430169] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.430173] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.430177] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1e2f690) 00:27:34.357 [2024-11-06 09:03:24.430184] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.357 [2024-11-06 09:03:24.430193] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91100, cid 0, qid 0 00:27:34.357 [2024-11-06 09:03:24.430404] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.357 [2024-11-06 09:03:24.430410] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.357 [2024-11-06 09:03:24.430414] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.430418] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91100) on tqpair=0x1e2f690 00:27:34.357 [2024-11-06 09:03:24.430422] nvme_ctrlr.c:3905:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:27:34.357 [2024-11-06 09:03:24.430430] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to reset admin queue (timeout 30000 ms) 00:27:34.357 [2024-11-06 09:03:24.430437] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to identify controller (no timeout) 00:27:34.357 [2024-11-06 09:03:24.430447] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for identify controller (timeout 30000 ms) 00:27:34.357 [2024-11-06 09:03:24.430456] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.430460] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1e2f690) 00:27:34.357 [2024-11-06 09:03:24.430468] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.357 [2024-11-06 09:03:24.430478] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91100, cid 0, qid 0 00:27:34.357 [2024-11-06 09:03:24.430686] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.357 [2024-11-06 09:03:24.430693] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.357 [2024-11-06 09:03:24.430697] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.430701] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1e2f690): datao=0, datal=4096, cccid=0 00:27:34.357 [2024-11-06 09:03:24.430706] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1e91100) on tqpair(0x1e2f690): expected_datao=0, payload_size=4096 00:27:34.357 [2024-11-06 09:03:24.430710] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.430723] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.357 [2024-11-06 09:03:24.430727] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.471914] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.621 [2024-11-06 09:03:24.471927] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.621 [2024-11-06 09:03:24.471931] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.471935] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91100) on tqpair=0x1e2f690 00:27:34.621 [2024-11-06 09:03:24.471944] nvme_ctrlr.c:2054:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] transport max_xfer_size 4294967295 00:27:34.621 [2024-11-06 09:03:24.471949] nvme_ctrlr.c:2058:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] MDTS max_xfer_size 131072 00:27:34.621 [2024-11-06 09:03:24.471954] nvme_ctrlr.c:2061:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] CNTLID 0x0001 00:27:34.621 [2024-11-06 09:03:24.471959] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] transport max_sges 16 00:27:34.621 [2024-11-06 09:03:24.471964] nvme_ctrlr.c:2100:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] fuses compare and write: 1 00:27:34.621 [2024-11-06 09:03:24.471969] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to configure AER (timeout 30000 ms) 00:27:34.621 [2024-11-06 09:03:24.471978] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for configure aer (timeout 30000 ms) 00:27:34.621 [2024-11-06 09:03:24.471985] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.471989] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.471993] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1e2f690) 00:27:34.621 [2024-11-06 09:03:24.472000] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:27:34.621 [2024-11-06 09:03:24.472013] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91100, cid 0, qid 0 00:27:34.621 [2024-11-06 09:03:24.472156] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.621 [2024-11-06 09:03:24.472163] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.621 [2024-11-06 09:03:24.472167] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.472171] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91100) on tqpair=0x1e2f690 00:27:34.621 [2024-11-06 09:03:24.472182] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.472185] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.472189] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1e2f690) 00:27:34.621 [2024-11-06 09:03:24.472195] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:27:34.621 [2024-11-06 09:03:24.472202] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.472206] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.472209] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x1e2f690) 00:27:34.621 [2024-11-06 09:03:24.472215] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:27:34.621 [2024-11-06 09:03:24.472221] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.472225] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.472229] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x1e2f690) 00:27:34.621 [2024-11-06 09:03:24.472235] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:27:34.621 [2024-11-06 09:03:24.472241] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.472244] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.472248] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1e2f690) 00:27:34.621 [2024-11-06 09:03:24.472254] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:27:34.621 [2024-11-06 09:03:24.472259] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to set keep alive timeout (timeout 30000 ms) 00:27:34.621 [2024-11-06 09:03:24.472267] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:27:34.621 [2024-11-06 09:03:24.472274] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.621 [2024-11-06 09:03:24.472277] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1e2f690) 00:27:34.621 [2024-11-06 09:03:24.472284] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.621 [2024-11-06 09:03:24.472296] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91100, cid 0, qid 0 00:27:34.621 [2024-11-06 09:03:24.472302] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91280, cid 1, qid 0 00:27:34.621 [2024-11-06 09:03:24.472306] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91400, cid 2, qid 0 00:27:34.621 [2024-11-06 09:03:24.472311] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91580, cid 3, qid 0 00:27:34.621 [2024-11-06 09:03:24.472316] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91700, cid 4, qid 0 00:27:34.621 [2024-11-06 09:03:24.472543] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.621 [2024-11-06 09:03:24.472550] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.621 [2024-11-06 09:03:24.472553] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.472557] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91700) on tqpair=0x1e2f690 00:27:34.622 [2024-11-06 09:03:24.472567] nvme_ctrlr.c:3023:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Sending keep alive every 5000000 us 00:27:34.622 [2024-11-06 09:03:24.472573] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] setting state to ready (no timeout) 00:27:34.622 [2024-11-06 09:03:24.472584] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.472588] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1e2f690) 00:27:34.622 [2024-11-06 09:03:24.472594] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.622 [2024-11-06 09:03:24.472605] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91700, cid 4, qid 0 00:27:34.622 [2024-11-06 09:03:24.472821] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.622 [2024-11-06 09:03:24.472828] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.622 [2024-11-06 09:03:24.472832] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.472836] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1e2f690): datao=0, datal=4096, cccid=4 00:27:34.622 [2024-11-06 09:03:24.472841] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1e91700) on tqpair(0x1e2f690): expected_datao=0, payload_size=4096 00:27:34.622 [2024-11-06 09:03:24.472845] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.472866] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.472870] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.520752] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.622 [2024-11-06 09:03:24.520763] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.622 [2024-11-06 09:03:24.520766] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.520771] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91700) on tqpair=0x1e2f690 00:27:34.622 [2024-11-06 09:03:24.520783] nvme_ctrlr.c:4166:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Ctrlr already in ready state 00:27:34.622 [2024-11-06 09:03:24.520807] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.520811] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1e2f690) 00:27:34.622 [2024-11-06 09:03:24.520818] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.622 [2024-11-06 09:03:24.520826] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.520830] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.520833] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x1e2f690) 00:27:34.622 [2024-11-06 09:03:24.520839] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:27:34.622 [2024-11-06 09:03:24.520854] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91700, cid 4, qid 0 00:27:34.622 [2024-11-06 09:03:24.520860] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91880, cid 5, qid 0 00:27:34.622 [2024-11-06 09:03:24.521182] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.622 [2024-11-06 09:03:24.521188] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.622 [2024-11-06 09:03:24.521192] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.521196] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1e2f690): datao=0, datal=1024, cccid=4 00:27:34.622 [2024-11-06 09:03:24.521200] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1e91700) on tqpair(0x1e2f690): expected_datao=0, payload_size=1024 00:27:34.622 [2024-11-06 09:03:24.521205] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.521214] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.521218] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.521224] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.622 [2024-11-06 09:03:24.521230] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.622 [2024-11-06 09:03:24.521233] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.521237] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91880) on tqpair=0x1e2f690 00:27:34.622 [2024-11-06 09:03:24.563754] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.622 [2024-11-06 09:03:24.563763] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.622 [2024-11-06 09:03:24.563767] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.563771] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91700) on tqpair=0x1e2f690 00:27:34.622 [2024-11-06 09:03:24.563782] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.563786] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1e2f690) 00:27:34.622 [2024-11-06 09:03:24.563793] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:02ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.622 [2024-11-06 09:03:24.563808] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91700, cid 4, qid 0 00:27:34.622 [2024-11-06 09:03:24.564003] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.622 [2024-11-06 09:03:24.564010] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.622 [2024-11-06 09:03:24.564013] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.564017] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1e2f690): datao=0, datal=3072, cccid=4 00:27:34.622 [2024-11-06 09:03:24.564022] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1e91700) on tqpair(0x1e2f690): expected_datao=0, payload_size=3072 00:27:34.622 [2024-11-06 09:03:24.564026] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.564043] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.564047] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.604921] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.622 [2024-11-06 09:03:24.604930] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.622 [2024-11-06 09:03:24.604933] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.604937] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91700) on tqpair=0x1e2f690 00:27:34.622 [2024-11-06 09:03:24.604946] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.604950] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1e2f690) 00:27:34.622 [2024-11-06 09:03:24.604957] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00010070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.622 [2024-11-06 09:03:24.604971] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91700, cid 4, qid 0 00:27:34.622 [2024-11-06 09:03:24.605224] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.622 [2024-11-06 09:03:24.605230] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.622 [2024-11-06 09:03:24.605234] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.605237] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1e2f690): datao=0, datal=8, cccid=4 00:27:34.622 [2024-11-06 09:03:24.605242] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1e91700) on tqpair(0x1e2f690): expected_datao=0, payload_size=8 00:27:34.622 [2024-11-06 09:03:24.605246] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.605253] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.605260] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.645949] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.622 [2024-11-06 09:03:24.645957] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.622 [2024-11-06 09:03:24.645961] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.622 [2024-11-06 09:03:24.645965] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91700) on tqpair=0x1e2f690 00:27:34.622 ===================================================== 00:27:34.622 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2014-08.org.nvmexpress.discovery 00:27:34.622 ===================================================== 00:27:34.622 Controller Capabilities/Features 00:27:34.622 ================================ 00:27:34.622 Vendor ID: 0000 00:27:34.622 Subsystem Vendor ID: 0000 00:27:34.622 Serial Number: .................... 00:27:34.622 Model Number: ........................................ 00:27:34.622 Firmware Version: 25.01 00:27:34.622 Recommended Arb Burst: 0 00:27:34.622 IEEE OUI Identifier: 00 00 00 00:27:34.622 Multi-path I/O 00:27:34.622 May have multiple subsystem ports: No 00:27:34.622 May have multiple controllers: No 00:27:34.622 Associated with SR-IOV VF: No 00:27:34.622 Max Data Transfer Size: 131072 00:27:34.622 Max Number of Namespaces: 0 00:27:34.622 Max Number of I/O Queues: 1024 00:27:34.622 NVMe Specification Version (VS): 1.3 00:27:34.622 NVMe Specification Version (Identify): 1.3 00:27:34.622 Maximum Queue Entries: 128 00:27:34.622 Contiguous Queues Required: Yes 00:27:34.622 Arbitration Mechanisms Supported 00:27:34.622 Weighted Round Robin: Not Supported 00:27:34.622 Vendor Specific: Not Supported 00:27:34.622 Reset Timeout: 15000 ms 00:27:34.622 Doorbell Stride: 4 bytes 00:27:34.622 NVM Subsystem Reset: Not Supported 00:27:34.622 Command Sets Supported 00:27:34.622 NVM Command Set: Supported 00:27:34.622 Boot Partition: Not Supported 00:27:34.622 Memory Page Size Minimum: 4096 bytes 00:27:34.622 Memory Page Size Maximum: 4096 bytes 00:27:34.622 Persistent Memory Region: Not Supported 00:27:34.622 Optional Asynchronous Events Supported 00:27:34.622 Namespace Attribute Notices: Not Supported 00:27:34.622 Firmware Activation Notices: Not Supported 00:27:34.622 ANA Change Notices: Not Supported 00:27:34.622 PLE Aggregate Log Change Notices: Not Supported 00:27:34.622 LBA Status Info Alert Notices: Not Supported 00:27:34.622 EGE Aggregate Log Change Notices: Not Supported 00:27:34.622 Normal NVM Subsystem Shutdown event: Not Supported 00:27:34.622 Zone Descriptor Change Notices: Not Supported 00:27:34.622 Discovery Log Change Notices: Supported 00:27:34.622 Controller Attributes 00:27:34.622 128-bit Host Identifier: Not Supported 00:27:34.622 Non-Operational Permissive Mode: Not Supported 00:27:34.622 NVM Sets: Not Supported 00:27:34.623 Read Recovery Levels: Not Supported 00:27:34.623 Endurance Groups: Not Supported 00:27:34.623 Predictable Latency Mode: Not Supported 00:27:34.623 Traffic Based Keep ALive: Not Supported 00:27:34.623 Namespace Granularity: Not Supported 00:27:34.623 SQ Associations: Not Supported 00:27:34.623 UUID List: Not Supported 00:27:34.623 Multi-Domain Subsystem: Not Supported 00:27:34.623 Fixed Capacity Management: Not Supported 00:27:34.623 Variable Capacity Management: Not Supported 00:27:34.623 Delete Endurance Group: Not Supported 00:27:34.623 Delete NVM Set: Not Supported 00:27:34.623 Extended LBA Formats Supported: Not Supported 00:27:34.623 Flexible Data Placement Supported: Not Supported 00:27:34.623 00:27:34.623 Controller Memory Buffer Support 00:27:34.623 ================================ 00:27:34.623 Supported: No 00:27:34.623 00:27:34.623 Persistent Memory Region Support 00:27:34.623 ================================ 00:27:34.623 Supported: No 00:27:34.623 00:27:34.623 Admin Command Set Attributes 00:27:34.623 ============================ 00:27:34.623 Security Send/Receive: Not Supported 00:27:34.623 Format NVM: Not Supported 00:27:34.623 Firmware Activate/Download: Not Supported 00:27:34.623 Namespace Management: Not Supported 00:27:34.623 Device Self-Test: Not Supported 00:27:34.623 Directives: Not Supported 00:27:34.623 NVMe-MI: Not Supported 00:27:34.623 Virtualization Management: Not Supported 00:27:34.623 Doorbell Buffer Config: Not Supported 00:27:34.623 Get LBA Status Capability: Not Supported 00:27:34.623 Command & Feature Lockdown Capability: Not Supported 00:27:34.623 Abort Command Limit: 1 00:27:34.623 Async Event Request Limit: 4 00:27:34.623 Number of Firmware Slots: N/A 00:27:34.623 Firmware Slot 1 Read-Only: N/A 00:27:34.623 Firmware Activation Without Reset: N/A 00:27:34.623 Multiple Update Detection Support: N/A 00:27:34.623 Firmware Update Granularity: No Information Provided 00:27:34.623 Per-Namespace SMART Log: No 00:27:34.623 Asymmetric Namespace Access Log Page: Not Supported 00:27:34.623 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:27:34.623 Command Effects Log Page: Not Supported 00:27:34.623 Get Log Page Extended Data: Supported 00:27:34.623 Telemetry Log Pages: Not Supported 00:27:34.623 Persistent Event Log Pages: Not Supported 00:27:34.623 Supported Log Pages Log Page: May Support 00:27:34.623 Commands Supported & Effects Log Page: Not Supported 00:27:34.623 Feature Identifiers & Effects Log Page:May Support 00:27:34.623 NVMe-MI Commands & Effects Log Page: May Support 00:27:34.623 Data Area 4 for Telemetry Log: Not Supported 00:27:34.623 Error Log Page Entries Supported: 128 00:27:34.623 Keep Alive: Not Supported 00:27:34.623 00:27:34.623 NVM Command Set Attributes 00:27:34.623 ========================== 00:27:34.623 Submission Queue Entry Size 00:27:34.623 Max: 1 00:27:34.623 Min: 1 00:27:34.623 Completion Queue Entry Size 00:27:34.623 Max: 1 00:27:34.623 Min: 1 00:27:34.623 Number of Namespaces: 0 00:27:34.623 Compare Command: Not Supported 00:27:34.623 Write Uncorrectable Command: Not Supported 00:27:34.623 Dataset Management Command: Not Supported 00:27:34.623 Write Zeroes Command: Not Supported 00:27:34.623 Set Features Save Field: Not Supported 00:27:34.623 Reservations: Not Supported 00:27:34.623 Timestamp: Not Supported 00:27:34.623 Copy: Not Supported 00:27:34.623 Volatile Write Cache: Not Present 00:27:34.623 Atomic Write Unit (Normal): 1 00:27:34.623 Atomic Write Unit (PFail): 1 00:27:34.623 Atomic Compare & Write Unit: 1 00:27:34.623 Fused Compare & Write: Supported 00:27:34.623 Scatter-Gather List 00:27:34.623 SGL Command Set: Supported 00:27:34.623 SGL Keyed: Supported 00:27:34.623 SGL Bit Bucket Descriptor: Not Supported 00:27:34.623 SGL Metadata Pointer: Not Supported 00:27:34.623 Oversized SGL: Not Supported 00:27:34.623 SGL Metadata Address: Not Supported 00:27:34.623 SGL Offset: Supported 00:27:34.623 Transport SGL Data Block: Not Supported 00:27:34.623 Replay Protected Memory Block: Not Supported 00:27:34.623 00:27:34.623 Firmware Slot Information 00:27:34.623 ========================= 00:27:34.623 Active slot: 0 00:27:34.623 00:27:34.623 00:27:34.623 Error Log 00:27:34.623 ========= 00:27:34.623 00:27:34.623 Active Namespaces 00:27:34.623 ================= 00:27:34.623 Discovery Log Page 00:27:34.623 ================== 00:27:34.623 Generation Counter: 2 00:27:34.623 Number of Records: 2 00:27:34.623 Record Format: 0 00:27:34.623 00:27:34.623 Discovery Log Entry 0 00:27:34.623 ---------------------- 00:27:34.623 Transport Type: 3 (TCP) 00:27:34.623 Address Family: 1 (IPv4) 00:27:34.623 Subsystem Type: 3 (Current Discovery Subsystem) 00:27:34.623 Entry Flags: 00:27:34.623 Duplicate Returned Information: 1 00:27:34.623 Explicit Persistent Connection Support for Discovery: 1 00:27:34.623 Transport Requirements: 00:27:34.623 Secure Channel: Not Required 00:27:34.623 Port ID: 0 (0x0000) 00:27:34.623 Controller ID: 65535 (0xffff) 00:27:34.623 Admin Max SQ Size: 128 00:27:34.623 Transport Service Identifier: 4420 00:27:34.623 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:27:34.623 Transport Address: 10.0.0.2 00:27:34.623 Discovery Log Entry 1 00:27:34.623 ---------------------- 00:27:34.623 Transport Type: 3 (TCP) 00:27:34.623 Address Family: 1 (IPv4) 00:27:34.623 Subsystem Type: 2 (NVM Subsystem) 00:27:34.623 Entry Flags: 00:27:34.623 Duplicate Returned Information: 0 00:27:34.623 Explicit Persistent Connection Support for Discovery: 0 00:27:34.623 Transport Requirements: 00:27:34.623 Secure Channel: Not Required 00:27:34.623 Port ID: 0 (0x0000) 00:27:34.623 Controller ID: 65535 (0xffff) 00:27:34.623 Admin Max SQ Size: 128 00:27:34.623 Transport Service Identifier: 4420 00:27:34.623 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:cnode1 00:27:34.623 Transport Address: 10.0.0.2 [2024-11-06 09:03:24.646049] nvme_ctrlr.c:4363:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] Prepare to destruct SSD 00:27:34.623 [2024-11-06 09:03:24.646061] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91100) on tqpair=0x1e2f690 00:27:34.623 [2024-11-06 09:03:24.646067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:34.623 [2024-11-06 09:03:24.646073] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91280) on tqpair=0x1e2f690 00:27:34.623 [2024-11-06 09:03:24.646078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:34.623 [2024-11-06 09:03:24.646083] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91400) on tqpair=0x1e2f690 00:27:34.623 [2024-11-06 09:03:24.646088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:34.623 [2024-11-06 09:03:24.646093] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91580) on tqpair=0x1e2f690 00:27:34.623 [2024-11-06 09:03:24.646097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:34.623 [2024-11-06 09:03:24.646106] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.623 [2024-11-06 09:03:24.646110] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.623 [2024-11-06 09:03:24.646113] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1e2f690) 00:27:34.623 [2024-11-06 09:03:24.646121] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.623 [2024-11-06 09:03:24.646134] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91580, cid 3, qid 0 00:27:34.623 [2024-11-06 09:03:24.646405] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.623 [2024-11-06 09:03:24.646411] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.623 [2024-11-06 09:03:24.646415] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.623 [2024-11-06 09:03:24.646419] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91580) on tqpair=0x1e2f690 00:27:34.623 [2024-11-06 09:03:24.646428] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.623 [2024-11-06 09:03:24.646432] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.623 [2024-11-06 09:03:24.646436] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1e2f690) 00:27:34.623 [2024-11-06 09:03:24.646443] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.623 [2024-11-06 09:03:24.646456] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91580, cid 3, qid 0 00:27:34.623 [2024-11-06 09:03:24.646705] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.623 [2024-11-06 09:03:24.646711] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.623 [2024-11-06 09:03:24.646715] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.623 [2024-11-06 09:03:24.646719] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91580) on tqpair=0x1e2f690 00:27:34.623 [2024-11-06 09:03:24.646724] nvme_ctrlr.c:1124:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] RTD3E = 0 us 00:27:34.623 [2024-11-06 09:03:24.646729] nvme_ctrlr.c:1127:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] shutdown timeout = 10000 ms 00:27:34.623 [2024-11-06 09:03:24.646741] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.623 [2024-11-06 09:03:24.646749] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.623 [2024-11-06 09:03:24.646753] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1e2f690) 00:27:34.624 [2024-11-06 09:03:24.646760] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.624 [2024-11-06 09:03:24.646770] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91580, cid 3, qid 0 00:27:34.624 [2024-11-06 09:03:24.647006] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.624 [2024-11-06 09:03:24.647012] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.624 [2024-11-06 09:03:24.647016] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.647020] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91580) on tqpair=0x1e2f690 00:27:34.624 [2024-11-06 09:03:24.647029] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.647033] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.647037] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1e2f690) 00:27:34.624 [2024-11-06 09:03:24.647044] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.624 [2024-11-06 09:03:24.647054] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91580, cid 3, qid 0 00:27:34.624 [2024-11-06 09:03:24.647220] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.624 [2024-11-06 09:03:24.647226] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.624 [2024-11-06 09:03:24.647230] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.647234] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91580) on tqpair=0x1e2f690 00:27:34.624 [2024-11-06 09:03:24.647243] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.647247] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.647251] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1e2f690) 00:27:34.624 [2024-11-06 09:03:24.647258] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.624 [2024-11-06 09:03:24.647268] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91580, cid 3, qid 0 00:27:34.624 [2024-11-06 09:03:24.647532] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.624 [2024-11-06 09:03:24.647538] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.624 [2024-11-06 09:03:24.647541] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.647545] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91580) on tqpair=0x1e2f690 00:27:34.624 [2024-11-06 09:03:24.647555] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.647559] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.647562] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1e2f690) 00:27:34.624 [2024-11-06 09:03:24.647569] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.624 [2024-11-06 09:03:24.647579] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91580, cid 3, qid 0 00:27:34.624 [2024-11-06 09:03:24.651752] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.624 [2024-11-06 09:03:24.651767] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.624 [2024-11-06 09:03:24.651771] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.651775] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91580) on tqpair=0x1e2f690 00:27:34.624 [2024-11-06 09:03:24.651784] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.651788] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.651794] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1e2f690) 00:27:34.624 [2024-11-06 09:03:24.651801] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.624 [2024-11-06 09:03:24.651813] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1e91580, cid 3, qid 0 00:27:34.624 [2024-11-06 09:03:24.651994] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.624 [2024-11-06 09:03:24.652000] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.624 [2024-11-06 09:03:24.652004] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.624 [2024-11-06 09:03:24.652008] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1e91580) on tqpair=0x1e2f690 00:27:34.624 [2024-11-06 09:03:24.652015] nvme_ctrlr.c:1246:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery, 1] shutdown complete in 5 milliseconds 00:27:34.624 00:27:34.624 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -L all 00:27:34.624 [2024-11-06 09:03:24.698694] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:27:34.624 [2024-11-06 09:03:24.698743] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid1985600 ] 00:27:34.889 [2024-11-06 09:03:24.752779] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 0] setting state to connect adminq (no timeout) 00:27:34.889 [2024-11-06 09:03:24.752835] nvme_tcp.c:2238:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:27:34.889 [2024-11-06 09:03:24.752840] nvme_tcp.c:2242:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:27:34.889 [2024-11-06 09:03:24.752854] nvme_tcp.c:2263:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:27:34.889 [2024-11-06 09:03:24.752862] sock.c: 373:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:27:34.889 [2024-11-06 09:03:24.756964] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 0] setting state to wait for connect adminq (no timeout) 00:27:34.889 [2024-11-06 09:03:24.756992] nvme_tcp.c:1455:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x1d1a690 0 00:27:34.890 [2024-11-06 09:03:24.764752] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:27:34.890 [2024-11-06 09:03:24.764765] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:27:34.890 [2024-11-06 09:03:24.764769] nvme_tcp.c:1501:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:27:34.890 [2024-11-06 09:03:24.764773] nvme_tcp.c:1502:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:27:34.890 [2024-11-06 09:03:24.764800] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.764805] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.764809] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1d1a690) 00:27:34.890 [2024-11-06 09:03:24.764821] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:27:34.890 [2024-11-06 09:03:24.764839] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c100, cid 0, qid 0 00:27:34.890 [2024-11-06 09:03:24.772756] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.890 [2024-11-06 09:03:24.772765] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.890 [2024-11-06 09:03:24.772769] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.772777] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c100) on tqpair=0x1d1a690 00:27:34.890 [2024-11-06 09:03:24.772786] nvme_fabric.c: 621:nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:27:34.890 [2024-11-06 09:03:24.772792] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read vs (no timeout) 00:27:34.890 [2024-11-06 09:03:24.772798] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read vs wait for vs (no timeout) 00:27:34.890 [2024-11-06 09:03:24.772810] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.772814] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.772817] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1d1a690) 00:27:34.890 [2024-11-06 09:03:24.772825] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.890 [2024-11-06 09:03:24.772838] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c100, cid 0, qid 0 00:27:34.890 [2024-11-06 09:03:24.772990] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.890 [2024-11-06 09:03:24.772997] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.890 [2024-11-06 09:03:24.773001] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773004] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c100) on tqpair=0x1d1a690 00:27:34.890 [2024-11-06 09:03:24.773009] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read cap (no timeout) 00:27:34.890 [2024-11-06 09:03:24.773017] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to read cap wait for cap (no timeout) 00:27:34.890 [2024-11-06 09:03:24.773024] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773028] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773031] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1d1a690) 00:27:34.890 [2024-11-06 09:03:24.773038] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.890 [2024-11-06 09:03:24.773049] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c100, cid 0, qid 0 00:27:34.890 [2024-11-06 09:03:24.773258] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.890 [2024-11-06 09:03:24.773265] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.890 [2024-11-06 09:03:24.773268] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773272] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c100) on tqpair=0x1d1a690 00:27:34.890 [2024-11-06 09:03:24.773277] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to check en (no timeout) 00:27:34.890 [2024-11-06 09:03:24.773285] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to check en wait for cc (timeout 15000 ms) 00:27:34.890 [2024-11-06 09:03:24.773292] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773295] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773299] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1d1a690) 00:27:34.890 [2024-11-06 09:03:24.773306] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.890 [2024-11-06 09:03:24.773316] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c100, cid 0, qid 0 00:27:34.890 [2024-11-06 09:03:24.773518] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.890 [2024-11-06 09:03:24.773524] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.890 [2024-11-06 09:03:24.773528] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773532] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c100) on tqpair=0x1d1a690 00:27:34.890 [2024-11-06 09:03:24.773539] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:27:34.890 [2024-11-06 09:03:24.773551] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773555] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773559] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1d1a690) 00:27:34.890 [2024-11-06 09:03:24.773566] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.890 [2024-11-06 09:03:24.773576] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c100, cid 0, qid 0 00:27:34.890 [2024-11-06 09:03:24.773784] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.890 [2024-11-06 09:03:24.773791] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.890 [2024-11-06 09:03:24.773794] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773798] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c100) on tqpair=0x1d1a690 00:27:34.890 [2024-11-06 09:03:24.773803] nvme_ctrlr.c:3870:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] CC.EN = 0 && CSTS.RDY = 0 00:27:34.890 [2024-11-06 09:03:24.773808] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to controller is disabled (timeout 15000 ms) 00:27:34.890 [2024-11-06 09:03:24.773816] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:27:34.890 [2024-11-06 09:03:24.773921] nvme_ctrlr.c:4068:nvme_ctrlr_process_init: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Setting CC.EN = 1 00:27:34.890 [2024-11-06 09:03:24.773926] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:27:34.890 [2024-11-06 09:03:24.773933] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773937] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.773941] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1d1a690) 00:27:34.890 [2024-11-06 09:03:24.773948] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.890 [2024-11-06 09:03:24.773958] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c100, cid 0, qid 0 00:27:34.890 [2024-11-06 09:03:24.774132] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.890 [2024-11-06 09:03:24.774139] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.890 [2024-11-06 09:03:24.774142] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.774146] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c100) on tqpair=0x1d1a690 00:27:34.890 [2024-11-06 09:03:24.774151] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:27:34.890 [2024-11-06 09:03:24.774160] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.774164] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.774168] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1d1a690) 00:27:34.890 [2024-11-06 09:03:24.774174] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.890 [2024-11-06 09:03:24.774185] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c100, cid 0, qid 0 00:27:34.890 [2024-11-06 09:03:24.774373] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.890 [2024-11-06 09:03:24.774380] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.890 [2024-11-06 09:03:24.774383] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.774389] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c100) on tqpair=0x1d1a690 00:27:34.890 [2024-11-06 09:03:24.774394] nvme_ctrlr.c:3905:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:27:34.890 [2024-11-06 09:03:24.774399] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to reset admin queue (timeout 30000 ms) 00:27:34.890 [2024-11-06 09:03:24.774406] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify controller (no timeout) 00:27:34.890 [2024-11-06 09:03:24.774418] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify controller (timeout 30000 ms) 00:27:34.890 [2024-11-06 09:03:24.774427] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.890 [2024-11-06 09:03:24.774430] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1d1a690) 00:27:34.890 [2024-11-06 09:03:24.774437] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.890 [2024-11-06 09:03:24.774448] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c100, cid 0, qid 0 00:27:34.890 [2024-11-06 09:03:24.774667] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.890 [2024-11-06 09:03:24.774674] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.891 [2024-11-06 09:03:24.774678] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.774682] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1d1a690): datao=0, datal=4096, cccid=0 00:27:34.891 [2024-11-06 09:03:24.774687] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1d7c100) on tqpair(0x1d1a690): expected_datao=0, payload_size=4096 00:27:34.891 [2024-11-06 09:03:24.774691] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.774706] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.774711] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.814918] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.891 [2024-11-06 09:03:24.814928] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.891 [2024-11-06 09:03:24.814931] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.814935] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c100) on tqpair=0x1d1a690 00:27:34.891 [2024-11-06 09:03:24.814943] nvme_ctrlr.c:2054:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] transport max_xfer_size 4294967295 00:27:34.891 [2024-11-06 09:03:24.814948] nvme_ctrlr.c:2058:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] MDTS max_xfer_size 131072 00:27:34.891 [2024-11-06 09:03:24.814952] nvme_ctrlr.c:2061:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] CNTLID 0x0001 00:27:34.891 [2024-11-06 09:03:24.814957] nvme_ctrlr.c:2085:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] transport max_sges 16 00:27:34.891 [2024-11-06 09:03:24.814962] nvme_ctrlr.c:2100:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] fuses compare and write: 1 00:27:34.891 [2024-11-06 09:03:24.814967] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to configure AER (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.814975] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for configure aer (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.814981] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.814985] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.814989] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1d1a690) 00:27:34.891 [2024-11-06 09:03:24.814996] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:27:34.891 [2024-11-06 09:03:24.815012] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c100, cid 0, qid 0 00:27:34.891 [2024-11-06 09:03:24.815165] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.891 [2024-11-06 09:03:24.815172] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.891 [2024-11-06 09:03:24.815175] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815179] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c100) on tqpair=0x1d1a690 00:27:34.891 [2024-11-06 09:03:24.815191] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815196] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815199] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1d1a690) 00:27:34.891 [2024-11-06 09:03:24.815206] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:27:34.891 [2024-11-06 09:03:24.815212] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815216] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815220] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x1d1a690) 00:27:34.891 [2024-11-06 09:03:24.815226] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:27:34.891 [2024-11-06 09:03:24.815232] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815237] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815241] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x1d1a690) 00:27:34.891 [2024-11-06 09:03:24.815248] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:27:34.891 [2024-11-06 09:03:24.815254] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815258] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815262] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1d1a690) 00:27:34.891 [2024-11-06 09:03:24.815267] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:27:34.891 [2024-11-06 09:03:24.815273] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set keep alive timeout (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.815282] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.815288] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815292] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1d1a690) 00:27:34.891 [2024-11-06 09:03:24.815298] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.891 [2024-11-06 09:03:24.815310] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c100, cid 0, qid 0 00:27:34.891 [2024-11-06 09:03:24.815316] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c280, cid 1, qid 0 00:27:34.891 [2024-11-06 09:03:24.815321] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c400, cid 2, qid 0 00:27:34.891 [2024-11-06 09:03:24.815326] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c580, cid 3, qid 0 00:27:34.891 [2024-11-06 09:03:24.815331] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c700, cid 4, qid 0 00:27:34.891 [2024-11-06 09:03:24.815520] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.891 [2024-11-06 09:03:24.815528] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.891 [2024-11-06 09:03:24.815532] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815538] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c700) on tqpair=0x1d1a690 00:27:34.891 [2024-11-06 09:03:24.815545] nvme_ctrlr.c:3023:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Sending keep alive every 5000000 us 00:27:34.891 [2024-11-06 09:03:24.815550] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify controller iocs specific (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.815559] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set number of queues (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.815565] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for set number of queues (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.815573] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815578] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815581] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1d1a690) 00:27:34.891 [2024-11-06 09:03:24.815588] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:4 cdw10:00000007 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:27:34.891 [2024-11-06 09:03:24.815598] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c700, cid 4, qid 0 00:27:34.891 [2024-11-06 09:03:24.815765] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.891 [2024-11-06 09:03:24.815772] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.891 [2024-11-06 09:03:24.815775] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815779] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c700) on tqpair=0x1d1a690 00:27:34.891 [2024-11-06 09:03:24.815843] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify active ns (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.815853] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify active ns (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.815861] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.815865] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1d1a690) 00:27:34.891 [2024-11-06 09:03:24.815871] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.891 [2024-11-06 09:03:24.815882] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c700, cid 4, qid 0 00:27:34.891 [2024-11-06 09:03:24.816053] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.891 [2024-11-06 09:03:24.816060] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.891 [2024-11-06 09:03:24.816063] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.816067] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1d1a690): datao=0, datal=4096, cccid=4 00:27:34.891 [2024-11-06 09:03:24.816071] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1d7c700) on tqpair(0x1d1a690): expected_datao=0, payload_size=4096 00:27:34.891 [2024-11-06 09:03:24.816076] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.816092] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.816097] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.816308] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.891 [2024-11-06 09:03:24.816314] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.891 [2024-11-06 09:03:24.816318] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.816322] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c700) on tqpair=0x1d1a690 00:27:34.891 [2024-11-06 09:03:24.816331] nvme_ctrlr.c:4699:spdk_nvme_ctrlr_get_ns: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Namespace 1 was added 00:27:34.891 [2024-11-06 09:03:24.816341] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify ns (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.816350] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify ns (timeout 30000 ms) 00:27:34.891 [2024-11-06 09:03:24.816357] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.891 [2024-11-06 09:03:24.816361] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1d1a690) 00:27:34.891 [2024-11-06 09:03:24.816368] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000000 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.891 [2024-11-06 09:03:24.816378] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c700, cid 4, qid 0 00:27:34.892 [2024-11-06 09:03:24.816608] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.892 [2024-11-06 09:03:24.816614] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.892 [2024-11-06 09:03:24.816618] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.816621] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1d1a690): datao=0, datal=4096, cccid=4 00:27:34.892 [2024-11-06 09:03:24.816626] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1d7c700) on tqpair(0x1d1a690): expected_datao=0, payload_size=4096 00:27:34.892 [2024-11-06 09:03:24.816630] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.816646] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.816650] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.820755] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.892 [2024-11-06 09:03:24.820763] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.892 [2024-11-06 09:03:24.820767] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.820771] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c700) on tqpair=0x1d1a690 00:27:34.892 [2024-11-06 09:03:24.820783] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify namespace id descriptors (timeout 30000 ms) 00:27:34.892 [2024-11-06 09:03:24.820793] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:27:34.892 [2024-11-06 09:03:24.820800] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.820804] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1d1a690) 00:27:34.892 [2024-11-06 09:03:24.820811] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.892 [2024-11-06 09:03:24.820822] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c700, cid 4, qid 0 00:27:34.892 [2024-11-06 09:03:24.821002] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.892 [2024-11-06 09:03:24.821008] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.892 [2024-11-06 09:03:24.821012] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821016] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1d1a690): datao=0, datal=4096, cccid=4 00:27:34.892 [2024-11-06 09:03:24.821020] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1d7c700) on tqpair(0x1d1a690): expected_datao=0, payload_size=4096 00:27:34.892 [2024-11-06 09:03:24.821024] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821031] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821035] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821201] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.892 [2024-11-06 09:03:24.821207] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.892 [2024-11-06 09:03:24.821214] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821218] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c700) on tqpair=0x1d1a690 00:27:34.892 [2024-11-06 09:03:24.821225] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to identify ns iocs specific (timeout 30000 ms) 00:27:34.892 [2024-11-06 09:03:24.821233] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set supported log pages (timeout 30000 ms) 00:27:34.892 [2024-11-06 09:03:24.821241] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set supported features (timeout 30000 ms) 00:27:34.892 [2024-11-06 09:03:24.821247] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set host behavior support feature (timeout 30000 ms) 00:27:34.892 [2024-11-06 09:03:24.821252] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set doorbell buffer config (timeout 30000 ms) 00:27:34.892 [2024-11-06 09:03:24.821257] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to set host ID (timeout 30000 ms) 00:27:34.892 [2024-11-06 09:03:24.821263] nvme_ctrlr.c:3111:nvme_ctrlr_set_host_id: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] NVMe-oF transport - not sending Set Features - Host ID 00:27:34.892 [2024-11-06 09:03:24.821267] nvme_ctrlr.c:1534:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to transport ready (timeout 30000 ms) 00:27:34.892 [2024-11-06 09:03:24.821273] nvme_ctrlr.c:1540:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] setting state to ready (no timeout) 00:27:34.892 [2024-11-06 09:03:24.821286] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821290] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1d1a690) 00:27:34.892 [2024-11-06 09:03:24.821297] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:4 cdw10:00000001 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.892 [2024-11-06 09:03:24.821304] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821308] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821311] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x1d1a690) 00:27:34.892 [2024-11-06 09:03:24.821317] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:27:34.892 [2024-11-06 09:03:24.821330] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c700, cid 4, qid 0 00:27:34.892 [2024-11-06 09:03:24.821336] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c880, cid 5, qid 0 00:27:34.892 [2024-11-06 09:03:24.821528] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.892 [2024-11-06 09:03:24.821534] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.892 [2024-11-06 09:03:24.821538] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821542] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c700) on tqpair=0x1d1a690 00:27:34.892 [2024-11-06 09:03:24.821549] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.892 [2024-11-06 09:03:24.821555] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.892 [2024-11-06 09:03:24.821558] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821562] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c880) on tqpair=0x1d1a690 00:27:34.892 [2024-11-06 09:03:24.821571] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821575] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x1d1a690) 00:27:34.892 [2024-11-06 09:03:24.821581] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:5 cdw10:00000002 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.892 [2024-11-06 09:03:24.821591] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c880, cid 5, qid 0 00:27:34.892 [2024-11-06 09:03:24.821755] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.892 [2024-11-06 09:03:24.821762] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.892 [2024-11-06 09:03:24.821766] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821769] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c880) on tqpair=0x1d1a690 00:27:34.892 [2024-11-06 09:03:24.821778] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.821782] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x1d1a690) 00:27:34.892 [2024-11-06 09:03:24.821789] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:5 cdw10:00000004 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.892 [2024-11-06 09:03:24.821799] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c880, cid 5, qid 0 00:27:34.892 [2024-11-06 09:03:24.821992] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.892 [2024-11-06 09:03:24.821998] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.892 [2024-11-06 09:03:24.822002] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.822006] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c880) on tqpair=0x1d1a690 00:27:34.892 [2024-11-06 09:03:24.822015] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.822019] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x1d1a690) 00:27:34.892 [2024-11-06 09:03:24.822025] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:5 cdw10:00000007 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.892 [2024-11-06 09:03:24.822035] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c880, cid 5, qid 0 00:27:34.892 [2024-11-06 09:03:24.822263] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.892 [2024-11-06 09:03:24.822269] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.892 [2024-11-06 09:03:24.822273] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.822277] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c880) on tqpair=0x1d1a690 00:27:34.892 [2024-11-06 09:03:24.822290] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.822294] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x1d1a690) 00:27:34.892 [2024-11-06 09:03:24.822301] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:5 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.892 [2024-11-06 09:03:24.822308] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.822312] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1d1a690) 00:27:34.892 [2024-11-06 09:03:24.822318] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:ffffffff cdw10:007f0002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.892 [2024-11-06 09:03:24.822326] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.822329] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=6 on tqpair(0x1d1a690) 00:27:34.892 [2024-11-06 09:03:24.822335] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:6 nsid:ffffffff cdw10:007f0003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.892 [2024-11-06 09:03:24.822345] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.892 [2024-11-06 09:03:24.822349] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x1d1a690) 00:27:34.892 [2024-11-06 09:03:24.822355] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:7 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.892 [2024-11-06 09:03:24.822366] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c880, cid 5, qid 0 00:27:34.893 [2024-11-06 09:03:24.822373] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c700, cid 4, qid 0 00:27:34.893 [2024-11-06 09:03:24.822378] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7ca00, cid 6, qid 0 00:27:34.893 [2024-11-06 09:03:24.822383] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7cb80, cid 7, qid 0 00:27:34.893 [2024-11-06 09:03:24.822634] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.893 [2024-11-06 09:03:24.822641] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.893 [2024-11-06 09:03:24.822644] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822648] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1d1a690): datao=0, datal=8192, cccid=5 00:27:34.893 [2024-11-06 09:03:24.822652] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1d7c880) on tqpair(0x1d1a690): expected_datao=0, payload_size=8192 00:27:34.893 [2024-11-06 09:03:24.822657] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822731] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822736] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822741] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.893 [2024-11-06 09:03:24.822751] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.893 [2024-11-06 09:03:24.822755] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822758] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1d1a690): datao=0, datal=512, cccid=4 00:27:34.893 [2024-11-06 09:03:24.822763] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1d7c700) on tqpair(0x1d1a690): expected_datao=0, payload_size=512 00:27:34.893 [2024-11-06 09:03:24.822767] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822774] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822777] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822783] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.893 [2024-11-06 09:03:24.822789] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.893 [2024-11-06 09:03:24.822792] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822796] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1d1a690): datao=0, datal=512, cccid=6 00:27:34.893 [2024-11-06 09:03:24.822800] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1d7ca00) on tqpair(0x1d1a690): expected_datao=0, payload_size=512 00:27:34.893 [2024-11-06 09:03:24.822804] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822811] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822814] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822820] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:27:34.893 [2024-11-06 09:03:24.822826] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:27:34.893 [2024-11-06 09:03:24.822829] nvme_tcp.c:1619:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822833] nvme_tcp.c:1620:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1d1a690): datao=0, datal=4096, cccid=7 00:27:34.893 [2024-11-06 09:03:24.822837] nvme_tcp.c:1631:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1d7cb80) on tqpair(0x1d1a690): expected_datao=0, payload_size=4096 00:27:34.893 [2024-11-06 09:03:24.822841] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822853] nvme_tcp.c:1421:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822856] nvme_tcp.c:1255:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822867] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.893 [2024-11-06 09:03:24.822873] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.893 [2024-11-06 09:03:24.822877] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822882] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c880) on tqpair=0x1d1a690 00:27:34.893 [2024-11-06 09:03:24.822896] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.893 [2024-11-06 09:03:24.822902] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.893 [2024-11-06 09:03:24.822906] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822909] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c700) on tqpair=0x1d1a690 00:27:34.893 [2024-11-06 09:03:24.822919] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.893 [2024-11-06 09:03:24.822925] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.893 [2024-11-06 09:03:24.822929] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822933] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7ca00) on tqpair=0x1d1a690 00:27:34.893 [2024-11-06 09:03:24.822940] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.893 [2024-11-06 09:03:24.822945] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.893 [2024-11-06 09:03:24.822949] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.893 [2024-11-06 09:03:24.822953] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7cb80) on tqpair=0x1d1a690 00:27:34.893 ===================================================== 00:27:34.893 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:27:34.893 ===================================================== 00:27:34.893 Controller Capabilities/Features 00:27:34.893 ================================ 00:27:34.893 Vendor ID: 8086 00:27:34.893 Subsystem Vendor ID: 8086 00:27:34.893 Serial Number: SPDK00000000000001 00:27:34.893 Model Number: SPDK bdev Controller 00:27:34.893 Firmware Version: 25.01 00:27:34.893 Recommended Arb Burst: 6 00:27:34.893 IEEE OUI Identifier: e4 d2 5c 00:27:34.893 Multi-path I/O 00:27:34.893 May have multiple subsystem ports: Yes 00:27:34.893 May have multiple controllers: Yes 00:27:34.893 Associated with SR-IOV VF: No 00:27:34.893 Max Data Transfer Size: 131072 00:27:34.893 Max Number of Namespaces: 32 00:27:34.893 Max Number of I/O Queues: 127 00:27:34.893 NVMe Specification Version (VS): 1.3 00:27:34.893 NVMe Specification Version (Identify): 1.3 00:27:34.893 Maximum Queue Entries: 128 00:27:34.893 Contiguous Queues Required: Yes 00:27:34.893 Arbitration Mechanisms Supported 00:27:34.893 Weighted Round Robin: Not Supported 00:27:34.893 Vendor Specific: Not Supported 00:27:34.893 Reset Timeout: 15000 ms 00:27:34.893 Doorbell Stride: 4 bytes 00:27:34.893 NVM Subsystem Reset: Not Supported 00:27:34.893 Command Sets Supported 00:27:34.893 NVM Command Set: Supported 00:27:34.893 Boot Partition: Not Supported 00:27:34.893 Memory Page Size Minimum: 4096 bytes 00:27:34.893 Memory Page Size Maximum: 4096 bytes 00:27:34.893 Persistent Memory Region: Not Supported 00:27:34.893 Optional Asynchronous Events Supported 00:27:34.893 Namespace Attribute Notices: Supported 00:27:34.893 Firmware Activation Notices: Not Supported 00:27:34.893 ANA Change Notices: Not Supported 00:27:34.893 PLE Aggregate Log Change Notices: Not Supported 00:27:34.893 LBA Status Info Alert Notices: Not Supported 00:27:34.893 EGE Aggregate Log Change Notices: Not Supported 00:27:34.893 Normal NVM Subsystem Shutdown event: Not Supported 00:27:34.893 Zone Descriptor Change Notices: Not Supported 00:27:34.893 Discovery Log Change Notices: Not Supported 00:27:34.893 Controller Attributes 00:27:34.893 128-bit Host Identifier: Supported 00:27:34.893 Non-Operational Permissive Mode: Not Supported 00:27:34.893 NVM Sets: Not Supported 00:27:34.893 Read Recovery Levels: Not Supported 00:27:34.893 Endurance Groups: Not Supported 00:27:34.893 Predictable Latency Mode: Not Supported 00:27:34.893 Traffic Based Keep ALive: Not Supported 00:27:34.893 Namespace Granularity: Not Supported 00:27:34.893 SQ Associations: Not Supported 00:27:34.893 UUID List: Not Supported 00:27:34.893 Multi-Domain Subsystem: Not Supported 00:27:34.893 Fixed Capacity Management: Not Supported 00:27:34.893 Variable Capacity Management: Not Supported 00:27:34.893 Delete Endurance Group: Not Supported 00:27:34.893 Delete NVM Set: Not Supported 00:27:34.893 Extended LBA Formats Supported: Not Supported 00:27:34.893 Flexible Data Placement Supported: Not Supported 00:27:34.893 00:27:34.893 Controller Memory Buffer Support 00:27:34.893 ================================ 00:27:34.893 Supported: No 00:27:34.893 00:27:34.893 Persistent Memory Region Support 00:27:34.893 ================================ 00:27:34.893 Supported: No 00:27:34.893 00:27:34.893 Admin Command Set Attributes 00:27:34.893 ============================ 00:27:34.893 Security Send/Receive: Not Supported 00:27:34.893 Format NVM: Not Supported 00:27:34.893 Firmware Activate/Download: Not Supported 00:27:34.893 Namespace Management: Not Supported 00:27:34.893 Device Self-Test: Not Supported 00:27:34.893 Directives: Not Supported 00:27:34.893 NVMe-MI: Not Supported 00:27:34.893 Virtualization Management: Not Supported 00:27:34.893 Doorbell Buffer Config: Not Supported 00:27:34.893 Get LBA Status Capability: Not Supported 00:27:34.893 Command & Feature Lockdown Capability: Not Supported 00:27:34.894 Abort Command Limit: 4 00:27:34.894 Async Event Request Limit: 4 00:27:34.894 Number of Firmware Slots: N/A 00:27:34.894 Firmware Slot 1 Read-Only: N/A 00:27:34.894 Firmware Activation Without Reset: N/A 00:27:34.894 Multiple Update Detection Support: N/A 00:27:34.894 Firmware Update Granularity: No Information Provided 00:27:34.894 Per-Namespace SMART Log: No 00:27:34.894 Asymmetric Namespace Access Log Page: Not Supported 00:27:34.894 Subsystem NQN: nqn.2016-06.io.spdk:cnode1 00:27:34.894 Command Effects Log Page: Supported 00:27:34.894 Get Log Page Extended Data: Supported 00:27:34.894 Telemetry Log Pages: Not Supported 00:27:34.894 Persistent Event Log Pages: Not Supported 00:27:34.894 Supported Log Pages Log Page: May Support 00:27:34.894 Commands Supported & Effects Log Page: Not Supported 00:27:34.894 Feature Identifiers & Effects Log Page:May Support 00:27:34.894 NVMe-MI Commands & Effects Log Page: May Support 00:27:34.894 Data Area 4 for Telemetry Log: Not Supported 00:27:34.894 Error Log Page Entries Supported: 128 00:27:34.894 Keep Alive: Supported 00:27:34.894 Keep Alive Granularity: 10000 ms 00:27:34.894 00:27:34.894 NVM Command Set Attributes 00:27:34.894 ========================== 00:27:34.894 Submission Queue Entry Size 00:27:34.894 Max: 64 00:27:34.894 Min: 64 00:27:34.894 Completion Queue Entry Size 00:27:34.894 Max: 16 00:27:34.894 Min: 16 00:27:34.894 Number of Namespaces: 32 00:27:34.894 Compare Command: Supported 00:27:34.894 Write Uncorrectable Command: Not Supported 00:27:34.894 Dataset Management Command: Supported 00:27:34.894 Write Zeroes Command: Supported 00:27:34.894 Set Features Save Field: Not Supported 00:27:34.894 Reservations: Supported 00:27:34.894 Timestamp: Not Supported 00:27:34.894 Copy: Supported 00:27:34.894 Volatile Write Cache: Present 00:27:34.894 Atomic Write Unit (Normal): 1 00:27:34.894 Atomic Write Unit (PFail): 1 00:27:34.894 Atomic Compare & Write Unit: 1 00:27:34.894 Fused Compare & Write: Supported 00:27:34.894 Scatter-Gather List 00:27:34.894 SGL Command Set: Supported 00:27:34.894 SGL Keyed: Supported 00:27:34.894 SGL Bit Bucket Descriptor: Not Supported 00:27:34.894 SGL Metadata Pointer: Not Supported 00:27:34.894 Oversized SGL: Not Supported 00:27:34.894 SGL Metadata Address: Not Supported 00:27:34.894 SGL Offset: Supported 00:27:34.894 Transport SGL Data Block: Not Supported 00:27:34.894 Replay Protected Memory Block: Not Supported 00:27:34.894 00:27:34.894 Firmware Slot Information 00:27:34.894 ========================= 00:27:34.894 Active slot: 1 00:27:34.894 Slot 1 Firmware Revision: 25.01 00:27:34.894 00:27:34.894 00:27:34.894 Commands Supported and Effects 00:27:34.894 ============================== 00:27:34.894 Admin Commands 00:27:34.894 -------------- 00:27:34.894 Get Log Page (02h): Supported 00:27:34.894 Identify (06h): Supported 00:27:34.894 Abort (08h): Supported 00:27:34.894 Set Features (09h): Supported 00:27:34.894 Get Features (0Ah): Supported 00:27:34.894 Asynchronous Event Request (0Ch): Supported 00:27:34.894 Keep Alive (18h): Supported 00:27:34.894 I/O Commands 00:27:34.894 ------------ 00:27:34.894 Flush (00h): Supported LBA-Change 00:27:34.894 Write (01h): Supported LBA-Change 00:27:34.894 Read (02h): Supported 00:27:34.894 Compare (05h): Supported 00:27:34.894 Write Zeroes (08h): Supported LBA-Change 00:27:34.894 Dataset Management (09h): Supported LBA-Change 00:27:34.894 Copy (19h): Supported LBA-Change 00:27:34.894 00:27:34.894 Error Log 00:27:34.894 ========= 00:27:34.894 00:27:34.894 Arbitration 00:27:34.894 =========== 00:27:34.894 Arbitration Burst: 1 00:27:34.894 00:27:34.894 Power Management 00:27:34.894 ================ 00:27:34.894 Number of Power States: 1 00:27:34.894 Current Power State: Power State #0 00:27:34.894 Power State #0: 00:27:34.894 Max Power: 0.00 W 00:27:34.894 Non-Operational State: Operational 00:27:34.894 Entry Latency: Not Reported 00:27:34.894 Exit Latency: Not Reported 00:27:34.894 Relative Read Throughput: 0 00:27:34.894 Relative Read Latency: 0 00:27:34.894 Relative Write Throughput: 0 00:27:34.894 Relative Write Latency: 0 00:27:34.894 Idle Power: Not Reported 00:27:34.894 Active Power: Not Reported 00:27:34.894 Non-Operational Permissive Mode: Not Supported 00:27:34.894 00:27:34.894 Health Information 00:27:34.894 ================== 00:27:34.894 Critical Warnings: 00:27:34.894 Available Spare Space: OK 00:27:34.894 Temperature: OK 00:27:34.894 Device Reliability: OK 00:27:34.894 Read Only: No 00:27:34.894 Volatile Memory Backup: OK 00:27:34.894 Current Temperature: 0 Kelvin (-273 Celsius) 00:27:34.894 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:27:34.894 Available Spare: 0% 00:27:34.894 Available Spare Threshold: 0% 00:27:34.894 Life Percentage Used:[2024-11-06 09:03:24.823046] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.894 [2024-11-06 09:03:24.823052] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x1d1a690) 00:27:34.894 [2024-11-06 09:03:24.823059] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:7 cdw10:00000005 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.894 [2024-11-06 09:03:24.823070] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7cb80, cid 7, qid 0 00:27:34.894 [2024-11-06 09:03:24.823229] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.894 [2024-11-06 09:03:24.823235] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.894 [2024-11-06 09:03:24.823239] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.894 [2024-11-06 09:03:24.823243] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7cb80) on tqpair=0x1d1a690 00:27:34.894 [2024-11-06 09:03:24.823273] nvme_ctrlr.c:4363:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] Prepare to destruct SSD 00:27:34.894 [2024-11-06 09:03:24.823282] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c100) on tqpair=0x1d1a690 00:27:34.894 [2024-11-06 09:03:24.823288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:34.894 [2024-11-06 09:03:24.823293] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c280) on tqpair=0x1d1a690 00:27:34.894 [2024-11-06 09:03:24.823298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:34.894 [2024-11-06 09:03:24.823303] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c400) on tqpair=0x1d1a690 00:27:34.894 [2024-11-06 09:03:24.823308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:34.894 [2024-11-06 09:03:24.823313] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c580) on tqpair=0x1d1a690 00:27:34.894 [2024-11-06 09:03:24.823317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:27:34.894 [2024-11-06 09:03:24.823325] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.894 [2024-11-06 09:03:24.823329] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.894 [2024-11-06 09:03:24.823333] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1d1a690) 00:27:34.894 [2024-11-06 09:03:24.823340] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.894 [2024-11-06 09:03:24.823351] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c580, cid 3, qid 0 00:27:34.895 [2024-11-06 09:03:24.823550] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.895 [2024-11-06 09:03:24.823557] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.895 [2024-11-06 09:03:24.823560] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.823564] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c580) on tqpair=0x1d1a690 00:27:34.895 [2024-11-06 09:03:24.823571] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.823575] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.823578] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1d1a690) 00:27:34.895 [2024-11-06 09:03:24.823585] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.895 [2024-11-06 09:03:24.823598] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c580, cid 3, qid 0 00:27:34.895 [2024-11-06 09:03:24.823815] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.895 [2024-11-06 09:03:24.823822] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.895 [2024-11-06 09:03:24.823825] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.823829] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c580) on tqpair=0x1d1a690 00:27:34.895 [2024-11-06 09:03:24.823834] nvme_ctrlr.c:1124:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] RTD3E = 0 us 00:27:34.895 [2024-11-06 09:03:24.823839] nvme_ctrlr.c:1127:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] shutdown timeout = 10000 ms 00:27:34.895 [2024-11-06 09:03:24.823848] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.823852] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.823856] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1d1a690) 00:27:34.895 [2024-11-06 09:03:24.823863] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.895 [2024-11-06 09:03:24.823873] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c580, cid 3, qid 0 00:27:34.895 [2024-11-06 09:03:24.824033] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.895 [2024-11-06 09:03:24.824039] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.895 [2024-11-06 09:03:24.824043] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824047] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c580) on tqpair=0x1d1a690 00:27:34.895 [2024-11-06 09:03:24.824056] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824060] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824064] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1d1a690) 00:27:34.895 [2024-11-06 09:03:24.824071] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.895 [2024-11-06 09:03:24.824081] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c580, cid 3, qid 0 00:27:34.895 [2024-11-06 09:03:24.824250] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.895 [2024-11-06 09:03:24.824256] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.895 [2024-11-06 09:03:24.824260] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824264] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c580) on tqpair=0x1d1a690 00:27:34.895 [2024-11-06 09:03:24.824273] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824277] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824281] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1d1a690) 00:27:34.895 [2024-11-06 09:03:24.824288] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.895 [2024-11-06 09:03:24.824301] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c580, cid 3, qid 0 00:27:34.895 [2024-11-06 09:03:24.824481] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.895 [2024-11-06 09:03:24.824488] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.895 [2024-11-06 09:03:24.824491] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824495] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c580) on tqpair=0x1d1a690 00:27:34.895 [2024-11-06 09:03:24.824505] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824508] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824512] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1d1a690) 00:27:34.895 [2024-11-06 09:03:24.824519] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.895 [2024-11-06 09:03:24.824529] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c580, cid 3, qid 0 00:27:34.895 [2024-11-06 09:03:24.824708] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.895 [2024-11-06 09:03:24.824714] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.895 [2024-11-06 09:03:24.824718] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824721] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c580) on tqpair=0x1d1a690 00:27:34.895 [2024-11-06 09:03:24.824731] nvme_tcp.c: 732:nvme_tcp_build_contig_request: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824735] nvme_tcp.c: 909:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.824738] nvme_tcp.c: 918:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1d1a690) 00:27:34.895 [2024-11-06 09:03:24.828751] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:27:34.895 [2024-11-06 09:03:24.828765] nvme_tcp.c: 883:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1d7c580, cid 3, qid 0 00:27:34.895 [2024-11-06 09:03:24.828933] nvme_tcp.c:1130:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:27:34.895 [2024-11-06 09:03:24.828939] nvme_tcp.c:1875:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:27:34.895 [2024-11-06 09:03:24.828942] nvme_tcp.c:1548:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:27:34.895 [2024-11-06 09:03:24.828946] nvme_tcp.c:1011:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1d7c580) on tqpair=0x1d1a690 00:27:34.895 [2024-11-06 09:03:24.828954] nvme_ctrlr.c:1246:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1, 1] shutdown complete in 5 milliseconds 00:27:34.895 0% 00:27:34.895 Data Units Read: 0 00:27:34.895 Data Units Written: 0 00:27:34.895 Host Read Commands: 0 00:27:34.895 Host Write Commands: 0 00:27:34.895 Controller Busy Time: 0 minutes 00:27:34.895 Power Cycles: 0 00:27:34.895 Power On Hours: 0 hours 00:27:34.895 Unsafe Shutdowns: 0 00:27:34.895 Unrecoverable Media Errors: 0 00:27:34.895 Lifetime Error Log Entries: 0 00:27:34.895 Warning Temperature Time: 0 minutes 00:27:34.895 Critical Temperature Time: 0 minutes 00:27:34.895 00:27:34.895 Number of Queues 00:27:34.895 ================ 00:27:34.895 Number of I/O Submission Queues: 127 00:27:34.895 Number of I/O Completion Queues: 127 00:27:34.895 00:27:34.895 Active Namespaces 00:27:34.895 ================= 00:27:34.895 Namespace ID:1 00:27:34.895 Error Recovery Timeout: Unlimited 00:27:34.895 Command Set Identifier: NVM (00h) 00:27:34.895 Deallocate: Supported 00:27:34.895 Deallocated/Unwritten Error: Not Supported 00:27:34.895 Deallocated Read Value: Unknown 00:27:34.895 Deallocate in Write Zeroes: Not Supported 00:27:34.895 Deallocated Guard Field: 0xFFFF 00:27:34.895 Flush: Supported 00:27:34.895 Reservation: Supported 00:27:34.895 Namespace Sharing Capabilities: Multiple Controllers 00:27:34.895 Size (in LBAs): 131072 (0GiB) 00:27:34.895 Capacity (in LBAs): 131072 (0GiB) 00:27:34.895 Utilization (in LBAs): 131072 (0GiB) 00:27:34.895 NGUID: ABCDEF0123456789ABCDEF0123456789 00:27:34.895 EUI64: ABCDEF0123456789 00:27:34.895 UUID: b44a0de2-7044-48ea-bf39-d1dbd43e4bc9 00:27:34.895 Thin Provisioning: Not Supported 00:27:34.895 Per-NS Atomic Units: Yes 00:27:34.895 Atomic Boundary Size (Normal): 0 00:27:34.895 Atomic Boundary Size (PFail): 0 00:27:34.895 Atomic Boundary Offset: 0 00:27:34.895 Maximum Single Source Range Length: 65535 00:27:34.895 Maximum Copy Length: 65535 00:27:34.895 Maximum Source Range Count: 1 00:27:34.895 NGUID/EUI64 Never Reused: No 00:27:34.895 Namespace Write Protected: No 00:27:34.895 Number of LBA Formats: 1 00:27:34.895 Current LBA Format: LBA Format #00 00:27:34.895 LBA Format #00: Data Size: 512 Metadata Size: 0 00:27:34.895 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@51 -- # sync 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@52 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@561 -- # xtrace_disable 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@54 -- # trap - SIGINT SIGTERM EXIT 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- host/identify.sh@56 -- # nvmftestfini 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@514 -- # nvmfcleanup 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@121 -- # sync 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@124 -- # set +e 00:27:34.895 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:34.896 rmmod nvme_tcp 00:27:34.896 rmmod nvme_fabrics 00:27:34.896 rmmod nvme_keyring 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@128 -- # set -e 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@129 -- # return 0 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@515 -- # '[' -n 1985250 ']' 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@516 -- # killprocess 1985250 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@950 -- # '[' -z 1985250 ']' 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@954 -- # kill -0 1985250 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@955 -- # uname 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1985250 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1985250' 00:27:34.896 killing process with pid 1985250 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@969 -- # kill 1985250 00:27:34.896 09:03:24 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@974 -- # wait 1985250 00:27:35.157 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:27:35.157 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:27:35.157 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:27:35.157 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@297 -- # iptr 00:27:35.158 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@789 -- # iptables-save 00:27:35.158 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:27:35.158 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@789 -- # iptables-restore 00:27:35.158 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:35.158 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:35.158 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:35.158 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:35.158 09:03:25 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:37.085 09:03:27 nvmf_tcp.nvmf_host.nvmf_identify -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:27:37.085 00:27:37.085 real 0m11.113s 00:27:37.085 user 0m8.675s 00:27:37.085 sys 0m5.646s 00:27:37.085 09:03:27 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@1126 -- # xtrace_disable 00:27:37.085 09:03:27 nvmf_tcp.nvmf_host.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:27:37.085 ************************************ 00:27:37.085 END TEST nvmf_identify 00:27:37.085 ************************************ 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@23 -- # run_test nvmf_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:27:37.346 ************************************ 00:27:37.346 START TEST nvmf_perf 00:27:37.346 ************************************ 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:27:37.346 * Looking for test storage... 00:27:37.346 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1689 -- # lcov --version 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@336 -- # IFS=.-: 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@336 -- # read -ra ver1 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@337 -- # IFS=.-: 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@337 -- # read -ra ver2 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@338 -- # local 'op=<' 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@340 -- # ver1_l=2 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@341 -- # ver2_l=1 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@344 -- # case "$op" in 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@345 -- # : 1 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@364 -- # (( v = 0 )) 00:27:37.346 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@365 -- # decimal 1 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@353 -- # local d=1 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@355 -- # echo 1 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@365 -- # ver1[v]=1 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@366 -- # decimal 2 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@353 -- # local d=2 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@355 -- # echo 2 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@366 -- # ver2[v]=2 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@368 -- # return 0 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:27:37.608 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:37.608 --rc genhtml_branch_coverage=1 00:27:37.608 --rc genhtml_function_coverage=1 00:27:37.608 --rc genhtml_legend=1 00:27:37.608 --rc geninfo_all_blocks=1 00:27:37.608 --rc geninfo_unexecuted_blocks=1 00:27:37.608 00:27:37.608 ' 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:27:37.608 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:37.608 --rc genhtml_branch_coverage=1 00:27:37.608 --rc genhtml_function_coverage=1 00:27:37.608 --rc genhtml_legend=1 00:27:37.608 --rc geninfo_all_blocks=1 00:27:37.608 --rc geninfo_unexecuted_blocks=1 00:27:37.608 00:27:37.608 ' 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:27:37.608 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:37.608 --rc genhtml_branch_coverage=1 00:27:37.608 --rc genhtml_function_coverage=1 00:27:37.608 --rc genhtml_legend=1 00:27:37.608 --rc geninfo_all_blocks=1 00:27:37.608 --rc geninfo_unexecuted_blocks=1 00:27:37.608 00:27:37.608 ' 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:27:37.608 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:27:37.608 --rc genhtml_branch_coverage=1 00:27:37.608 --rc genhtml_function_coverage=1 00:27:37.608 --rc genhtml_legend=1 00:27:37.608 --rc geninfo_all_blocks=1 00:27:37.608 --rc geninfo_unexecuted_blocks=1 00:27:37.608 00:27:37.608 ' 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@7 -- # uname -s 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:37.608 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@15 -- # shopt -s extglob 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@5 -- # export PATH 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@51 -- # : 0 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:27:37.609 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@12 -- # MALLOC_BDEV_SIZE=64 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@17 -- # nvmftestinit 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@474 -- # prepare_net_devs 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@436 -- # local -g is_hw=no 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@438 -- # remove_spdk_ns 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@309 -- # xtrace_disable 00:27:37.609 09:03:27 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@315 -- # pci_devs=() 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@315 -- # local -a pci_devs 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@317 -- # pci_drivers=() 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@319 -- # net_devs=() 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@319 -- # local -ga net_devs 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@320 -- # e810=() 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@320 -- # local -ga e810 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@321 -- # x722=() 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@321 -- # local -ga x722 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@322 -- # mlx=() 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@322 -- # local -ga mlx 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:27:45.750 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:27:45.750 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:45.750 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:27:45.751 Found net devices under 0000:4b:00.0: cvl_0_0 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:27:45.751 Found net devices under 0000:4b:00.1: cvl_0_1 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@440 -- # is_hw=yes 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:27:45.751 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:45.751 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.697 ms 00:27:45.751 00:27:45.751 --- 10.0.0.2 ping statistics --- 00:27:45.751 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:45.751 rtt min/avg/max/mdev = 0.697/0.697/0.697/0.000 ms 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:45.751 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:45.751 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.316 ms 00:27:45.751 00:27:45.751 --- 10.0.0.1 ping statistics --- 00:27:45.751 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:45.751 rtt min/avg/max/mdev = 0.316/0.316/0.316/0.000 ms 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@448 -- # return 0 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@18 -- # nvmfappstart -m 0xF 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@724 -- # xtrace_disable 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@507 -- # nvmfpid=1989838 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@508 -- # waitforlisten 1989838 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@831 -- # '[' -z 1989838 ']' 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@836 -- # local max_retries=100 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:45.751 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@840 -- # xtrace_disable 00:27:45.751 09:03:34 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:27:45.751 [2024-11-06 09:03:34.988553] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:27:45.751 [2024-11-06 09:03:34.988650] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:45.751 [2024-11-06 09:03:35.073225] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:27:45.751 [2024-11-06 09:03:35.116593] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:45.751 [2024-11-06 09:03:35.116631] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:45.751 [2024-11-06 09:03:35.116639] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:45.751 [2024-11-06 09:03:35.116646] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:45.751 [2024-11-06 09:03:35.116651] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:45.751 [2024-11-06 09:03:35.118280] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:27:45.751 [2024-11-06 09:03:35.118396] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:27:45.751 [2024-11-06 09:03:35.118554] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:27:45.751 [2024-11-06 09:03:35.118554] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:27:45.751 09:03:35 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:27:45.751 09:03:35 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@864 -- # return 0 00:27:45.751 09:03:35 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:27:45.751 09:03:35 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@730 -- # xtrace_disable 00:27:45.751 09:03:35 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:27:45.751 09:03:35 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:45.751 09:03:35 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:27:45.751 09:03:35 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:27:46.322 09:03:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_get_config bdev 00:27:46.322 09:03:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # jq -r '.[].params | select(.name=="Nvme0").traddr' 00:27:46.583 09:03:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@30 -- # local_nvme_trid=0000:65:00.0 00:27:46.583 09:03:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:27:46.844 09:03:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@31 -- # bdevs=' Malloc0' 00:27:46.844 09:03:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@33 -- # '[' -n 0000:65:00.0 ']' 00:27:46.844 09:03:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@34 -- # bdevs=' Malloc0 Nvme0n1' 00:27:46.844 09:03:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@37 -- # '[' tcp == rdma ']' 00:27:46.844 09:03:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:27:46.844 [2024-11-06 09:03:36.872777] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:46.844 09:03:36 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:27:47.104 09:03:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@45 -- # for bdev in $bdevs 00:27:47.104 09:03:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:27:47.365 09:03:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@45 -- # for bdev in $bdevs 00:27:47.365 09:03:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:27:47.365 09:03:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:27:47.625 [2024-11-06 09:03:37.595444] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:47.625 09:03:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:27:47.886 09:03:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@52 -- # '[' -n 0000:65:00.0 ']' 00:27:47.886 09:03:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@53 -- # perf_app -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:65:00.0' 00:27:47.886 09:03:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@21 -- # '[' 0 -eq 1 ']' 00:27:47.886 09:03:37 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:65:00.0' 00:27:49.270 Initializing NVMe Controllers 00:27:49.270 Attached to NVMe Controller at 0000:65:00.0 [144d:a80a] 00:27:49.270 Associating PCIE (0000:65:00.0) NSID 1 with lcore 0 00:27:49.270 Initialization complete. Launching workers. 00:27:49.271 ======================================================== 00:27:49.271 Latency(us) 00:27:49.271 Device Information : IOPS MiB/s Average min max 00:27:49.271 PCIE (0000:65:00.0) NSID 1 from core 0: 79523.91 310.64 401.67 13.26 6015.12 00:27:49.271 ======================================================== 00:27:49.271 Total : 79523.91 310.64 401.67 13.26 6015.12 00:27:49.271 00:27:49.271 09:03:39 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 1 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:27:50.653 Initializing NVMe Controllers 00:27:50.653 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:27:50.653 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:27:50.653 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:27:50.653 Initialization complete. Launching workers. 00:27:50.653 ======================================================== 00:27:50.653 Latency(us) 00:27:50.653 Device Information : IOPS MiB/s Average min max 00:27:50.653 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 122.68 0.48 8364.24 271.11 45631.40 00:27:50.653 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 74.80 0.29 13368.13 7954.86 53875.23 00:27:50.653 ======================================================== 00:27:50.653 Total : 197.48 0.77 10259.65 271.11 53875.23 00:27:50.653 00:27:50.653 09:03:40 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 32 -o 4096 -w randrw -M 50 -t 1 -HI -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:27:52.153 Initializing NVMe Controllers 00:27:52.153 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:27:52.153 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:27:52.153 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:27:52.153 Initialization complete. Launching workers. 00:27:52.153 ======================================================== 00:27:52.153 Latency(us) 00:27:52.153 Device Information : IOPS MiB/s Average min max 00:27:52.153 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 10262.29 40.09 3118.48 559.04 9021.67 00:27:52.153 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 3763.74 14.70 8515.04 6249.34 18235.33 00:27:52.153 ======================================================== 00:27:52.153 Total : 14026.03 54.79 4566.59 559.04 18235.33 00:27:52.153 00:27:52.153 09:03:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@59 -- # [[ e810 == \e\8\1\0 ]] 00:27:52.153 09:03:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@59 -- # [[ tcp == \r\d\m\a ]] 00:27:52.153 09:03:41 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -O 16384 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:27:54.698 Initializing NVMe Controllers 00:27:54.698 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:27:54.698 Controller IO queue size 128, less than required. 00:27:54.698 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:27:54.698 Controller IO queue size 128, less than required. 00:27:54.698 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:27:54.698 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:27:54.698 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:27:54.698 Initialization complete. Launching workers. 00:27:54.698 ======================================================== 00:27:54.698 Latency(us) 00:27:54.698 Device Information : IOPS MiB/s Average min max 00:27:54.698 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1924.59 481.15 66974.44 38266.00 105463.97 00:27:54.698 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 603.87 150.97 222320.47 62264.05 340643.64 00:27:54.698 ======================================================== 00:27:54.698 Total : 2528.46 632.11 104075.67 38266.00 340643.64 00:27:54.698 00:27:54.698 09:03:44 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 36964 -O 4096 -w randrw -M 50 -t 5 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0xf -P 4 00:27:54.698 No valid NVMe controllers or AIO or URING devices found 00:27:54.698 Initializing NVMe Controllers 00:27:54.698 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:27:54.698 Controller IO queue size 128, less than required. 00:27:54.698 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:27:54.698 WARNING: IO size 36964 (-o) is not a multiple of nsid 1 sector size 512. Removing this ns from test 00:27:54.698 Controller IO queue size 128, less than required. 00:27:54.698 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:27:54.698 WARNING: IO size 36964 (-o) is not a multiple of nsid 2 sector size 512. Removing this ns from test 00:27:54.698 WARNING: Some requested NVMe devices were skipped 00:27:54.698 09:03:44 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' --transport-stat 00:27:57.997 Initializing NVMe Controllers 00:27:57.997 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:27:57.997 Controller IO queue size 128, less than required. 00:27:57.997 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:27:57.998 Controller IO queue size 128, less than required. 00:27:57.998 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:27:57.998 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:27:57.998 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:27:57.998 Initialization complete. Launching workers. 00:27:57.998 00:27:57.998 ==================== 00:27:57.998 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 statistics: 00:27:57.998 TCP transport: 00:27:57.998 polls: 22098 00:27:57.998 idle_polls: 12323 00:27:57.998 sock_completions: 9775 00:27:57.998 nvme_completions: 6571 00:27:57.998 submitted_requests: 9802 00:27:57.998 queued_requests: 1 00:27:57.998 00:27:57.998 ==================== 00:27:57.998 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 statistics: 00:27:57.998 TCP transport: 00:27:57.998 polls: 21431 00:27:57.998 idle_polls: 11456 00:27:57.998 sock_completions: 9975 00:27:57.998 nvme_completions: 6669 00:27:57.998 submitted_requests: 10030 00:27:57.998 queued_requests: 1 00:27:57.998 ======================================================== 00:27:57.998 Latency(us) 00:27:57.998 Device Information : IOPS MiB/s Average min max 00:27:57.998 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1639.72 409.93 78860.23 50595.56 137159.76 00:27:57.998 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 1664.18 416.04 78036.92 33830.62 113777.30 00:27:57.998 ======================================================== 00:27:57.998 Total : 3303.90 825.97 78445.53 33830.62 137159.76 00:27:57.998 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@66 -- # sync 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@67 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@69 -- # '[' 0 -eq 1 ']' 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@112 -- # trap - SIGINT SIGTERM EXIT 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- host/perf.sh@114 -- # nvmftestfini 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@514 -- # nvmfcleanup 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@121 -- # sync 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@124 -- # set +e 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@125 -- # for i in {1..20} 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:27:57.998 rmmod nvme_tcp 00:27:57.998 rmmod nvme_fabrics 00:27:57.998 rmmod nvme_keyring 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@128 -- # set -e 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@129 -- # return 0 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@515 -- # '[' -n 1989838 ']' 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@516 -- # killprocess 1989838 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@950 -- # '[' -z 1989838 ']' 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@954 -- # kill -0 1989838 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@955 -- # uname 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1989838 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1989838' 00:27:57.998 killing process with pid 1989838 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@969 -- # kill 1989838 00:27:57.998 09:03:47 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@974 -- # wait 1989838 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@297 -- # iptr 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@789 -- # iptables-save 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@789 -- # iptables-restore 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:27:59.909 09:03:49 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:01.822 09:03:51 nvmf_tcp.nvmf_host.nvmf_perf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:28:01.822 00:28:01.823 real 0m24.517s 00:28:01.823 user 0m59.757s 00:28:01.823 sys 0m8.485s 00:28:01.823 09:03:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:28:01.823 09:03:51 nvmf_tcp.nvmf_host.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:28:01.823 ************************************ 00:28:01.823 END TEST nvmf_perf 00:28:01.823 ************************************ 00:28:01.823 09:03:51 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@24 -- # run_test nvmf_fio_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:28:01.823 09:03:51 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:28:01.823 09:03:51 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:28:01.823 09:03:51 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:28:01.823 ************************************ 00:28:01.823 START TEST nvmf_fio_host 00:28:01.823 ************************************ 00:28:01.823 09:03:51 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:28:02.085 * Looking for test storage... 00:28:02.085 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:28:02.085 09:03:51 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:28:02.085 09:03:51 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1689 -- # lcov --version 00:28:02.085 09:03:51 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@336 -- # IFS=.-: 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@336 -- # read -ra ver1 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@337 -- # IFS=.-: 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@337 -- # read -ra ver2 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@338 -- # local 'op=<' 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@340 -- # ver1_l=2 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@341 -- # ver2_l=1 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@344 -- # case "$op" in 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@345 -- # : 1 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@365 -- # decimal 1 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@353 -- # local d=1 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@355 -- # echo 1 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@365 -- # ver1[v]=1 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@366 -- # decimal 2 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@353 -- # local d=2 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@355 -- # echo 2 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@366 -- # ver2[v]=2 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@368 -- # return 0 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:28:02.085 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:02.085 --rc genhtml_branch_coverage=1 00:28:02.085 --rc genhtml_function_coverage=1 00:28:02.085 --rc genhtml_legend=1 00:28:02.085 --rc geninfo_all_blocks=1 00:28:02.085 --rc geninfo_unexecuted_blocks=1 00:28:02.085 00:28:02.085 ' 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:28:02.085 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:02.085 --rc genhtml_branch_coverage=1 00:28:02.085 --rc genhtml_function_coverage=1 00:28:02.085 --rc genhtml_legend=1 00:28:02.085 --rc geninfo_all_blocks=1 00:28:02.085 --rc geninfo_unexecuted_blocks=1 00:28:02.085 00:28:02.085 ' 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:28:02.085 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:02.085 --rc genhtml_branch_coverage=1 00:28:02.085 --rc genhtml_function_coverage=1 00:28:02.085 --rc genhtml_legend=1 00:28:02.085 --rc geninfo_all_blocks=1 00:28:02.085 --rc geninfo_unexecuted_blocks=1 00:28:02.085 00:28:02.085 ' 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:28:02.085 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:02.085 --rc genhtml_branch_coverage=1 00:28:02.085 --rc genhtml_function_coverage=1 00:28:02.085 --rc genhtml_legend=1 00:28:02.085 --rc geninfo_all_blocks=1 00:28:02.085 --rc geninfo_unexecuted_blocks=1 00:28:02.085 00:28:02.085 ' 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@15 -- # shopt -s extglob 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@5 -- # export PATH 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@7 -- # uname -s 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:28:02.085 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@15 -- # shopt -s extglob 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@5 -- # export PATH 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@51 -- # : 0 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:28:02.086 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@14 -- # nvmftestinit 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@474 -- # prepare_net_devs 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@436 -- # local -g is_hw=no 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@438 -- # remove_spdk_ns 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@309 -- # xtrace_disable 00:28:02.086 09:03:52 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:28:10.234 09:03:58 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:28:10.234 09:03:58 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@315 -- # pci_devs=() 00:28:10.234 09:03:58 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@315 -- # local -a pci_devs 00:28:10.234 09:03:58 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@316 -- # pci_net_devs=() 00:28:10.234 09:03:58 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:28:10.234 09:03:58 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@317 -- # pci_drivers=() 00:28:10.234 09:03:58 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@317 -- # local -A pci_drivers 00:28:10.234 09:03:58 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@319 -- # net_devs=() 00:28:10.234 09:03:58 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@319 -- # local -ga net_devs 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@320 -- # e810=() 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@320 -- # local -ga e810 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@321 -- # x722=() 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@321 -- # local -ga x722 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@322 -- # mlx=() 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@322 -- # local -ga mlx 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:28:10.234 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:28:10.234 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@416 -- # [[ up == up ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:28:10.234 Found net devices under 0000:4b:00.0: cvl_0_0 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@416 -- # [[ up == up ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:28:10.234 Found net devices under 0000:4b:00.1: cvl_0_1 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@440 -- # is_hw=yes 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:28:10.234 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:28:10.235 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:28:10.235 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.661 ms 00:28:10.235 00:28:10.235 --- 10.0.0.2 ping statistics --- 00:28:10.235 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:10.235 rtt min/avg/max/mdev = 0.661/0.661/0.661/0.000 ms 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:28:10.235 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:28:10.235 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.310 ms 00:28:10.235 00:28:10.235 --- 10.0.0.1 ping statistics --- 00:28:10.235 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:10.235 rtt min/avg/max/mdev = 0.310/0.310/0.310/0.000 ms 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@448 -- # return 0 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@16 -- # [[ y != y ]] 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@21 -- # timing_enter start_nvmf_tgt 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@724 -- # xtrace_disable 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@24 -- # nvmfpid=1996839 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@26 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@23 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@28 -- # waitforlisten 1996839 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@831 -- # '[' -z 1996839 ']' 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:10.235 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:10.235 09:03:59 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:28:10.235 [2024-11-06 09:03:59.398625] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:28:10.235 [2024-11-06 09:03:59.398693] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:10.235 [2024-11-06 09:03:59.481945] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:28:10.235 [2024-11-06 09:03:59.524390] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:28:10.235 [2024-11-06 09:03:59.524429] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:28:10.235 [2024-11-06 09:03:59.524438] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:28:10.235 [2024-11-06 09:03:59.524445] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:28:10.235 [2024-11-06 09:03:59.524451] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:28:10.235 [2024-11-06 09:03:59.526321] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:28:10.235 [2024-11-06 09:03:59.526440] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:28:10.235 [2024-11-06 09:03:59.526596] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:10.235 [2024-11-06 09:03:59.526596] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:28:10.235 09:04:00 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:10.235 09:04:00 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@864 -- # return 0 00:28:10.235 09:04:00 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:28:10.496 [2024-11-06 09:04:00.355483] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:28:10.496 09:04:00 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@30 -- # timing_exit start_nvmf_tgt 00:28:10.496 09:04:00 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@730 -- # xtrace_disable 00:28:10.496 09:04:00 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:28:10.496 09:04:00 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:28:10.756 Malloc1 00:28:10.756 09:04:00 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:28:10.756 09:04:00 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:28:11.017 09:04:00 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:11.278 [2024-11-06 09:04:01.150644] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@38 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@41 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1339 -- # local sanitizers 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1341 -- # shift 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local asan_lib= 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libasan 00:28:11.278 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:11.563 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:11.564 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:11.564 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:11.564 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:28:11.564 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:28:11.564 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:11.564 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:11.564 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:11.564 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:28:11.564 09:04:01 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:28:11.831 test: (g=0): rw=randrw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk, iodepth=128 00:28:11.831 fio-3.35 00:28:11.831 Starting 1 thread 00:28:14.370 00:28:14.370 test: (groupid=0, jobs=1): err= 0: pid=1997533: Wed Nov 6 09:04:04 2024 00:28:14.370 read: IOPS=12.9k, BW=50.3MiB/s (52.7MB/s)(101MiB/2005msec) 00:28:14.370 slat (usec): min=2, max=232, avg= 2.15, stdev= 2.04 00:28:14.370 clat (usec): min=3679, max=8884, avg=5465.54, stdev=902.14 00:28:14.370 lat (usec): min=3717, max=8886, avg=5467.69, stdev=902.17 00:28:14.370 clat percentiles (usec): 00:28:14.370 | 1.00th=[ 4359], 5.00th=[ 4621], 10.00th=[ 4752], 20.00th=[ 4883], 00:28:14.370 | 30.00th=[ 5014], 40.00th=[ 5080], 50.00th=[ 5211], 60.00th=[ 5276], 00:28:14.370 | 70.00th=[ 5407], 80.00th=[ 5604], 90.00th=[ 7177], 95.00th=[ 7635], 00:28:14.370 | 99.00th=[ 8160], 99.50th=[ 8356], 99.90th=[ 8717], 99.95th=[ 8717], 00:28:14.370 | 99.99th=[ 8848] 00:28:14.370 bw ( KiB/s): min=39264, max=55672, per=99.99%, avg=51484.00, stdev=8147.55, samples=4 00:28:14.370 iops : min= 9816, max=13918, avg=12871.00, stdev=2036.89, samples=4 00:28:14.370 write: IOPS=12.8k, BW=50.2MiB/s (52.6MB/s)(101MiB/2005msec); 0 zone resets 00:28:14.370 slat (usec): min=2, max=229, avg= 2.22, stdev= 1.59 00:28:14.370 clat (usec): min=2569, max=8094, avg=4420.21, stdev=745.85 00:28:14.370 lat (usec): min=2587, max=8096, avg=4422.43, stdev=745.92 00:28:14.370 clat percentiles (usec): 00:28:14.370 | 1.00th=[ 3490], 5.00th=[ 3720], 10.00th=[ 3818], 20.00th=[ 3949], 00:28:14.370 | 30.00th=[ 4015], 40.00th=[ 4113], 50.00th=[ 4178], 60.00th=[ 4293], 00:28:14.370 | 70.00th=[ 4359], 80.00th=[ 4555], 90.00th=[ 5866], 95.00th=[ 6194], 00:28:14.370 | 99.00th=[ 6652], 99.50th=[ 6783], 99.90th=[ 7177], 99.95th=[ 7373], 00:28:14.370 | 99.99th=[ 7832] 00:28:14.370 bw ( KiB/s): min=39816, max=55432, per=100.00%, avg=51396.00, stdev=7721.14, samples=4 00:28:14.370 iops : min= 9954, max=13858, avg=12849.00, stdev=1930.28, samples=4 00:28:14.370 lat (msec) : 4=13.23%, 10=86.77% 00:28:14.370 cpu : usr=73.50%, sys=25.10%, ctx=20, majf=0, minf=17 00:28:14.370 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=99.9% 00:28:14.370 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:14.370 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:28:14.370 issued rwts: total=25808,25759,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:14.370 latency : target=0, window=0, percentile=100.00%, depth=128 00:28:14.370 00:28:14.370 Run status group 0 (all jobs): 00:28:14.370 READ: bw=50.3MiB/s (52.7MB/s), 50.3MiB/s-50.3MiB/s (52.7MB/s-52.7MB/s), io=101MiB (106MB), run=2005-2005msec 00:28:14.370 WRITE: bw=50.2MiB/s (52.6MB/s), 50.2MiB/s-50.2MiB/s (52.6MB/s-52.6MB/s), io=101MiB (106MB), run=2005-2005msec 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@45 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1339 -- # local sanitizers 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1341 -- # shift 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local asan_lib= 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libasan 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:28:14.370 09:04:04 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:28:14.632 test: (g=0): rw=randrw, bs=(R) 16.0KiB-16.0KiB, (W) 16.0KiB-16.0KiB, (T) 16.0KiB-16.0KiB, ioengine=spdk, iodepth=128 00:28:14.632 fio-3.35 00:28:14.632 Starting 1 thread 00:28:17.177 00:28:17.177 test: (groupid=0, jobs=1): err= 0: pid=1998285: Wed Nov 6 09:04:06 2024 00:28:17.177 read: IOPS=9309, BW=145MiB/s (153MB/s)(292MiB/2009msec) 00:28:17.177 slat (usec): min=3, max=110, avg= 3.63, stdev= 1.68 00:28:17.177 clat (usec): min=1348, max=17663, avg=8359.13, stdev=2055.72 00:28:17.177 lat (usec): min=1352, max=17666, avg=8362.76, stdev=2055.90 00:28:17.177 clat percentiles (usec): 00:28:17.177 | 1.00th=[ 4228], 5.00th=[ 5211], 10.00th=[ 5800], 20.00th=[ 6587], 00:28:17.177 | 30.00th=[ 7177], 40.00th=[ 7701], 50.00th=[ 8225], 60.00th=[ 8848], 00:28:17.177 | 70.00th=[ 9372], 80.00th=[10159], 90.00th=[11076], 95.00th=[11731], 00:28:17.177 | 99.00th=[13566], 99.50th=[14353], 99.90th=[16581], 99.95th=[16909], 00:28:17.177 | 99.99th=[17695] 00:28:17.177 bw ( KiB/s): min=66912, max=81408, per=49.49%, avg=73712.00, stdev=6058.32, samples=4 00:28:17.177 iops : min= 4182, max= 5088, avg=4607.00, stdev=378.64, samples=4 00:28:17.177 write: IOPS=5230, BW=81.7MiB/s (85.7MB/s)(150MiB/1834msec); 0 zone resets 00:28:17.177 slat (usec): min=39, max=449, avg=41.15, stdev= 9.43 00:28:17.177 clat (usec): min=2029, max=16968, avg=9571.74, stdev=1655.17 00:28:17.177 lat (usec): min=2069, max=17102, avg=9612.89, stdev=1657.69 00:28:17.177 clat percentiles (usec): 00:28:17.177 | 1.00th=[ 6390], 5.00th=[ 7308], 10.00th=[ 7635], 20.00th=[ 8225], 00:28:17.177 | 30.00th=[ 8717], 40.00th=[ 8979], 50.00th=[ 9372], 60.00th=[ 9896], 00:28:17.177 | 70.00th=[10290], 80.00th=[10814], 90.00th=[11731], 95.00th=[12518], 00:28:17.177 | 99.00th=[14091], 99.50th=[14877], 99.90th=[16712], 99.95th=[16909], 00:28:17.177 | 99.99th=[16909] 00:28:17.177 bw ( KiB/s): min=68096, max=85312, per=91.34%, avg=76432.00, stdev=7078.95, samples=4 00:28:17.177 iops : min= 4256, max= 5332, avg=4777.00, stdev=442.43, samples=4 00:28:17.177 lat (msec) : 2=0.04%, 4=0.49%, 10=73.09%, 20=26.38% 00:28:17.177 cpu : usr=84.81%, sys=13.45%, ctx=16, majf=0, minf=23 00:28:17.177 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.4%, 32=0.8%, >=64=98.4% 00:28:17.177 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:17.177 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:28:17.177 issued rwts: total=18703,9592,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:17.177 latency : target=0, window=0, percentile=100.00%, depth=128 00:28:17.177 00:28:17.177 Run status group 0 (all jobs): 00:28:17.177 READ: bw=145MiB/s (153MB/s), 145MiB/s-145MiB/s (153MB/s-153MB/s), io=292MiB (306MB), run=2009-2009msec 00:28:17.177 WRITE: bw=81.7MiB/s (85.7MB/s), 81.7MiB/s-81.7MiB/s (85.7MB/s-85.7MB/s), io=150MiB (157MB), run=1834-1834msec 00:28:17.177 09:04:06 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@49 -- # '[' 0 -eq 1 ']' 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@85 -- # rm -f ./local-test-0-verify.state 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- host/fio.sh@86 -- # nvmftestfini 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@514 -- # nvmfcleanup 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@121 -- # sync 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@124 -- # set +e 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@125 -- # for i in {1..20} 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:28:17.177 rmmod nvme_tcp 00:28:17.177 rmmod nvme_fabrics 00:28:17.177 rmmod nvme_keyring 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@128 -- # set -e 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@129 -- # return 0 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@515 -- # '[' -n 1996839 ']' 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@516 -- # killprocess 1996839 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@950 -- # '[' -z 1996839 ']' 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@954 -- # kill -0 1996839 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@955 -- # uname 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:17.177 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 1996839 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@968 -- # echo 'killing process with pid 1996839' 00:28:17.438 killing process with pid 1996839 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@969 -- # kill 1996839 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@974 -- # wait 1996839 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@297 -- # iptr 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@789 -- # iptables-save 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@789 -- # iptables-restore 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@302 -- # remove_spdk_ns 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:17.438 09:04:07 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:19.985 09:04:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:28:19.985 00:28:19.985 real 0m17.643s 00:28:19.985 user 1m10.951s 00:28:19.985 sys 0m7.388s 00:28:19.985 09:04:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@1126 -- # xtrace_disable 00:28:19.985 09:04:09 nvmf_tcp.nvmf_host.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:28:19.985 ************************************ 00:28:19.985 END TEST nvmf_fio_host 00:28:19.985 ************************************ 00:28:19.985 09:04:09 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@25 -- # run_test nvmf_failover /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:28:19.985 09:04:09 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:28:19.985 09:04:09 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:28:19.985 09:04:09 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:28:19.985 ************************************ 00:28:19.985 START TEST nvmf_failover 00:28:19.985 ************************************ 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:28:19.986 * Looking for test storage... 00:28:19.986 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1689 -- # lcov --version 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@333 -- # local ver1 ver1_l 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@334 -- # local ver2 ver2_l 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@336 -- # IFS=.-: 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@336 -- # read -ra ver1 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@337 -- # IFS=.-: 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@337 -- # read -ra ver2 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@338 -- # local 'op=<' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@340 -- # ver1_l=2 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@341 -- # ver2_l=1 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@344 -- # case "$op" in 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@345 -- # : 1 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@364 -- # (( v = 0 )) 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@365 -- # decimal 1 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@353 -- # local d=1 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@355 -- # echo 1 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@365 -- # ver1[v]=1 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@366 -- # decimal 2 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@353 -- # local d=2 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@355 -- # echo 2 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@366 -- # ver2[v]=2 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@368 -- # return 0 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:28:19.986 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:19.986 --rc genhtml_branch_coverage=1 00:28:19.986 --rc genhtml_function_coverage=1 00:28:19.986 --rc genhtml_legend=1 00:28:19.986 --rc geninfo_all_blocks=1 00:28:19.986 --rc geninfo_unexecuted_blocks=1 00:28:19.986 00:28:19.986 ' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:28:19.986 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:19.986 --rc genhtml_branch_coverage=1 00:28:19.986 --rc genhtml_function_coverage=1 00:28:19.986 --rc genhtml_legend=1 00:28:19.986 --rc geninfo_all_blocks=1 00:28:19.986 --rc geninfo_unexecuted_blocks=1 00:28:19.986 00:28:19.986 ' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:28:19.986 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:19.986 --rc genhtml_branch_coverage=1 00:28:19.986 --rc genhtml_function_coverage=1 00:28:19.986 --rc genhtml_legend=1 00:28:19.986 --rc geninfo_all_blocks=1 00:28:19.986 --rc geninfo_unexecuted_blocks=1 00:28:19.986 00:28:19.986 ' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:28:19.986 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:19.986 --rc genhtml_branch_coverage=1 00:28:19.986 --rc genhtml_function_coverage=1 00:28:19.986 --rc genhtml_legend=1 00:28:19.986 --rc geninfo_all_blocks=1 00:28:19.986 --rc geninfo_unexecuted_blocks=1 00:28:19.986 00:28:19.986 ' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@7 -- # uname -s 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@15 -- # shopt -s extglob 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@5 -- # export PATH 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@51 -- # : 0 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:28:19.986 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@55 -- # have_pci_nics=0 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@11 -- # MALLOC_BDEV_SIZE=64 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:28:19.986 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@18 -- # nvmftestinit 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@474 -- # prepare_net_devs 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@436 -- # local -g is_hw=no 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@438 -- # remove_spdk_ns 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@309 -- # xtrace_disable 00:28:19.987 09:04:09 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:28:26.578 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:28:26.578 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@315 -- # pci_devs=() 00:28:26.578 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@315 -- # local -a pci_devs 00:28:26.578 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@316 -- # pci_net_devs=() 00:28:26.578 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:28:26.578 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@317 -- # pci_drivers=() 00:28:26.578 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@317 -- # local -A pci_drivers 00:28:26.578 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@319 -- # net_devs=() 00:28:26.578 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@319 -- # local -ga net_devs 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@320 -- # e810=() 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@320 -- # local -ga e810 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@321 -- # x722=() 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@321 -- # local -ga x722 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@322 -- # mlx=() 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@322 -- # local -ga mlx 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:28:26.840 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:28:26.840 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@416 -- # [[ up == up ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:28:26.840 Found net devices under 0000:4b:00.0: cvl_0_0 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@416 -- # [[ up == up ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:28:26.840 Found net devices under 0000:4b:00.1: cvl_0_1 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@440 -- # is_hw=yes 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:28:26.840 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:28:26.841 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:28:26.841 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:28:26.841 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:28:26.841 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:28:26.841 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:28:26.841 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:28:26.841 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:28:27.102 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:28:27.102 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.627 ms 00:28:27.102 00:28:27.102 --- 10.0.0.2 ping statistics --- 00:28:27.102 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:27.102 rtt min/avg/max/mdev = 0.627/0.627/0.627/0.000 ms 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:28:27.102 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:28:27.102 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.277 ms 00:28:27.102 00:28:27.102 --- 10.0.0.1 ping statistics --- 00:28:27.102 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:28:27.102 rtt min/avg/max/mdev = 0.277/0.277/0.277/0.000 ms 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@448 -- # return 0 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:28:27.102 09:04:16 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@20 -- # nvmfappstart -m 0xE 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@724 -- # xtrace_disable 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@507 -- # nvmfpid=2002865 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@508 -- # waitforlisten 2002865 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@831 -- # '[' -z 2002865 ']' 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:28:27.102 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:27.102 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:28:27.102 [2024-11-06 09:04:17.101152] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:28:27.102 [2024-11-06 09:04:17.101216] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:28:27.102 [2024-11-06 09:04:17.197765] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:28:27.363 [2024-11-06 09:04:17.233731] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:28:27.363 [2024-11-06 09:04:17.233772] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:28:27.363 [2024-11-06 09:04:17.233780] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:28:27.363 [2024-11-06 09:04:17.233787] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:28:27.363 [2024-11-06 09:04:17.233792] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:28:27.363 [2024-11-06 09:04:17.235311] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:28:27.363 [2024-11-06 09:04:17.235465] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:28:27.363 [2024-11-06 09:04:17.235466] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:28:27.934 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:27.934 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # return 0 00:28:27.934 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:28:27.934 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@730 -- # xtrace_disable 00:28:27.934 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:28:27.934 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:28:27.934 09:04:17 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:28:28.195 [2024-11-06 09:04:18.086185] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:28:28.195 09:04:18 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:28:28.195 Malloc0 00:28:28.456 09:04:18 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:28:28.456 09:04:18 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:28:28.717 09:04:18 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:28.978 [2024-11-06 09:04:18.836229] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:28.978 09:04:18 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:28:28.978 [2024-11-06 09:04:19.020718] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:28:28.978 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:28:29.239 [2024-11-06 09:04:19.205295] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:28:29.239 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@31 -- # bdevperf_pid=2003380 00:28:29.239 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 15 -f 00:28:29.239 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; cat $testdir/try.txt; rm -f $testdir/try.txt; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:28:29.239 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@34 -- # waitforlisten 2003380 /var/tmp/bdevperf.sock 00:28:29.239 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@831 -- # '[' -z 2003380 ']' 00:28:29.239 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:28:29.239 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:29.239 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:28:29.239 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:28:29.239 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:29.239 09:04:19 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:28:30.181 09:04:20 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:30.181 09:04:20 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # return 0 00:28:30.181 09:04:20 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:28:30.442 NVMe0n1 00:28:30.442 09:04:20 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:28:30.703 00:28:30.703 09:04:20 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@39 -- # run_test_pid=2003640 00:28:30.703 09:04:20 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:28:30.703 09:04:20 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@41 -- # sleep 1 00:28:31.645 09:04:21 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:31.906 [2024-11-06 09:04:21.821509] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.906 [2024-11-06 09:04:21.821552] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.906 [2024-11-06 09:04:21.821558] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.906 [2024-11-06 09:04:21.821563] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.906 [2024-11-06 09:04:21.821567] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.906 [2024-11-06 09:04:21.821572] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.906 [2024-11-06 09:04:21.821577] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.906 [2024-11-06 09:04:21.821582] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.906 [2024-11-06 09:04:21.821586] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.906 [2024-11-06 09:04:21.821591] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821595] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821600] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821604] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821609] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821613] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821617] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821622] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821627] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821631] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821635] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821640] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821645] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821649] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821654] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821659] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821663] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821668] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821678] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821682] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821687] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821691] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821695] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821700] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821705] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821709] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821714] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821718] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821723] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821727] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821731] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821736] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821740] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821751] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821755] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821760] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821764] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821769] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821773] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821778] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821782] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821787] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821791] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821796] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821800] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821804] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 [2024-11-06 09:04:21.821810] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f4ed0 is same with the state(6) to be set 00:28:31.907 09:04:21 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@45 -- # sleep 3 00:28:35.209 09:04:24 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:28:35.209 00:28:35.209 09:04:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:28:35.470 [2024-11-06 09:04:25.337658] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337693] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337700] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337705] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337710] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337715] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337720] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337724] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337730] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337735] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337740] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337745] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337754] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337759] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337763] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337768] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337772] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337777] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337783] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337788] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337793] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.470 [2024-11-06 09:04:25.337798] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337807] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337812] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337816] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337822] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337828] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337832] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337837] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337841] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337846] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337851] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337857] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337862] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337867] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337871] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337876] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 [2024-11-06 09:04:25.337880] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f5cf0 is same with the state(6) to be set 00:28:35.471 09:04:25 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@50 -- # sleep 3 00:28:38.773 09:04:28 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:38.773 [2024-11-06 09:04:28.524516] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:38.773 09:04:28 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@55 -- # sleep 1 00:28:39.716 09:04:29 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:28:39.716 [2024-11-06 09:04:29.718600] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718642] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718649] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718653] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718659] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718663] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718668] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718678] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718683] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718687] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718692] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718697] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718701] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718706] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718710] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718715] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718719] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718724] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718728] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718733] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718737] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718742] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718751] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718756] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718761] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718766] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718770] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718775] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718779] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 [2024-11-06 09:04:29.718784] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x9f6bf0 is same with the state(6) to be set 00:28:39.716 09:04:29 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@59 -- # wait 2003640 00:28:46.318 { 00:28:46.318 "results": [ 00:28:46.318 { 00:28:46.318 "job": "NVMe0n1", 00:28:46.318 "core_mask": "0x1", 00:28:46.318 "workload": "verify", 00:28:46.318 "status": "finished", 00:28:46.318 "verify_range": { 00:28:46.318 "start": 0, 00:28:46.318 "length": 16384 00:28:46.318 }, 00:28:46.318 "queue_depth": 128, 00:28:46.318 "io_size": 4096, 00:28:46.318 "runtime": 15.005156, 00:28:46.318 "iops": 11110.847498020014, 00:28:46.318 "mibps": 43.40174803914068, 00:28:46.318 "io_failed": 10052, 00:28:46.318 "io_timeout": 0, 00:28:46.318 "avg_latency_us": 10837.174852729315, 00:28:46.318 "min_latency_us": 570.0266666666666, 00:28:46.318 "max_latency_us": 29272.746666666666 00:28:46.318 } 00:28:46.318 ], 00:28:46.318 "core_count": 1 00:28:46.318 } 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@61 -- # killprocess 2003380 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@950 -- # '[' -z 2003380 ']' 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # kill -0 2003380 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # uname 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2003380 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2003380' 00:28:46.318 killing process with pid 2003380 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@969 -- # kill 2003380 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@974 -- # wait 2003380 00:28:46.318 09:04:35 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@63 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:28:46.318 [2024-11-06 09:04:19.292381] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:28:46.318 [2024-11-06 09:04:19.292443] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2003380 ] 00:28:46.318 [2024-11-06 09:04:19.364781] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:46.318 [2024-11-06 09:04:19.400397] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:46.318 Running I/O for 15 seconds... 00:28:46.318 11813.00 IOPS, 46.14 MiB/s [2024-11-06T08:04:36.431Z] [2024-11-06 09:04:21.822923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:101592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.822957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.822974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:101600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.822983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.822993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:101608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:101616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:101624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:101632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:101640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:101648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823094] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:101656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:101664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:101672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:101680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:101688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:101696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:101704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:101712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:101720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:101728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:101736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:101744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:101752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.318 [2024-11-06 09:04:21.823309] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.318 [2024-11-06 09:04:21.823318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:101760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:101768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:101776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:101784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:101792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:101800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:101808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:101816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:101824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:101832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:101840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:101848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:101856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:101864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:101872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:101880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823593] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:101888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:101896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:101904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:101912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:101920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:101928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:101936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:101944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:101952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:101960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:101968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:101976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:101984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:101992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:102000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:102008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:102016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:102024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:102032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823915] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:102040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:102048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:102056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:102064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.319 [2024-11-06 09:04:21.823971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.319 [2024-11-06 09:04:21.823980] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:102072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.320 [2024-11-06 09:04:21.823987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.823996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:102080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.320 [2024-11-06 09:04:21.824005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:102088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.320 [2024-11-06 09:04:21.824021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:102096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.320 [2024-11-06 09:04:21.824038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824058] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101088 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824105] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.320 [2024-11-06 09:04:21.824114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824123] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.320 [2024-11-06 09:04:21.824131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824139] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.320 [2024-11-06 09:04:21.824147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824155] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.320 [2024-11-06 09:04:21.824162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824169] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x17f1d70 is same with the state(6) to be set 00:28:46.320 [2024-11-06 09:04:21.824327] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824335] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101096 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824358] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824363] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101104 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824386] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824392] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101112 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824419] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824426] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101120 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824451] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824457] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101128 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824482] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824489] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101136 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824514] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824521] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101144 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824543] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824549] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101152 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824570] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824576] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101160 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824597] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824602] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101168 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824624] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824631] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101176 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824651] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824657] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101184 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824677] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824683] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101192 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824704] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824709] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824715] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101200 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824731] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824736] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102104 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824761] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824767] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101208 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824788] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824794] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101216 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.320 [2024-11-06 09:04:21.824815] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.320 [2024-11-06 09:04:21.824820] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.320 [2024-11-06 09:04:21.824827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101224 len:8 PRP1 0x0 PRP2 0x0 00:28:46.320 [2024-11-06 09:04:21.824834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.824843] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.824848] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.824855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101232 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.824862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.824869] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.824875] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.824881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101240 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.824888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.824895] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.824901] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.824907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101248 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.824914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.824921] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.824926] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.824932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101256 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.824939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.824947] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.824957] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.824963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101264 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.824970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.824978] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.824983] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.824989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101272 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.824996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825004] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825009] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101280 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.825022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825029] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825035] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101288 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.825050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825057] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825063] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101296 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.825075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825084] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825089] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101304 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.825103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825110] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825116] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101312 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.825130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825138] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825143] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101320 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.825157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825165] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825171] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101328 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.825184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825191] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825198] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101336 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.825213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825221] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825226] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101344 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.825240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825249] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825254] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101352 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.825270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.825279] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.825284] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.825290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101360 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.835932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.835966] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.835974] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.835982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101368 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.835991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.835999] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.836005] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.836011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101376 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.836018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.836026] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.836031] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.836038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101384 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.836045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.836053] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.836059] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.836065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101392 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.836072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.836080] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.836085] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.836091] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101400 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.836098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.836106] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.836112] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.321 [2024-11-06 09:04:21.836118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101408 len:8 PRP1 0x0 PRP2 0x0 00:28:46.321 [2024-11-06 09:04:21.836125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.321 [2024-11-06 09:04:21.836137] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.321 [2024-11-06 09:04:21.836142] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101416 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836163] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836169] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101424 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836190] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836196] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101432 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836217] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836222] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836229] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101440 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836244] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836250] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101448 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836271] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836276] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101456 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836297] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836303] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101464 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836324] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836329] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101472 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836344] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836352] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836357] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101480 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836378] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836383] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101488 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836404] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836410] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101496 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836430] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836436] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101504 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836457] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836463] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101512 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836484] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836490] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836496] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101520 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836510] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836516] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101528 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836536] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836542] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101536 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836565] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836570] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101544 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836591] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836597] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101552 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836618] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836623] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101560 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836644] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836649] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101568 len:8 PRP1 0x0 PRP2 0x0 00:28:46.322 [2024-11-06 09:04:21.836662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.322 [2024-11-06 09:04:21.836669] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.322 [2024-11-06 09:04:21.836675] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.322 [2024-11-06 09:04:21.836682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101576 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836696] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836702] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101584 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836723] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836728] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101592 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836758] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836765] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101600 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836785] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836791] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836797] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101608 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836812] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836817] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101616 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836838] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836844] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101624 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836865] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836871] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101632 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836891] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836896] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101640 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836917] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836922] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836928] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101648 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836944] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836949] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101656 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836971] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.836977] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.836983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101664 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.836990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.836998] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837003] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101672 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837023] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837029] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101680 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837049] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837055] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101688 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837075] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837081] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101696 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837102] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837107] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101704 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837129] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837134] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101712 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837155] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837161] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101720 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837184] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837190] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101728 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837211] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837216] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101736 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837237] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837243] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101744 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837264] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837270] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101752 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.323 [2024-11-06 09:04:21.837291] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.323 [2024-11-06 09:04:21.837296] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.323 [2024-11-06 09:04:21.837302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101760 len:8 PRP1 0x0 PRP2 0x0 00:28:46.323 [2024-11-06 09:04:21.837309] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837317] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837323] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101768 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837344] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837351] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837357] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101776 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837372] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837381] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101784 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837401] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837407] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101792 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837428] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837434] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101800 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837455] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837461] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101808 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837482] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837487] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101816 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837508] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837514] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101824 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837534] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837540] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101832 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837561] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837566] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101840 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837588] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837594] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101848 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837614] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837620] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101856 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837641] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837647] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101864 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837667] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837673] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101872 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.837685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.837693] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.837700] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.837706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101880 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.845349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.845379] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.845388] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.845397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101888 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.845405] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.845413] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.845419] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.845425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101896 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.845432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.845440] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.845445] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.845452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101904 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.845463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.845471] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.845476] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.845482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101912 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.845490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.845497] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.845502] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.845508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101920 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.845516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.845524] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.845529] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.845535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101928 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.845542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.845550] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.845556] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.845563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101936 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.845570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.845578] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.845583] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.845589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101944 len:8 PRP1 0x0 PRP2 0x0 00:28:46.324 [2024-11-06 09:04:21.845597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.324 [2024-11-06 09:04:21.845604] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.324 [2024-11-06 09:04:21.845610] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.324 [2024-11-06 09:04:21.845616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101952 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845630] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845636] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101960 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845657] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845663] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101968 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845685] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845691] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101976 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845712] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845717] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101984 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845738] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845744] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:101992 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845774] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845780] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102000 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845800] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845806] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102008 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845826] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845832] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102016 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845853] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845858] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102024 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845881] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845886] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102032 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845907] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845913] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102040 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845933] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845939] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102048 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845959] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845965] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102056 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.845978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.845986] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.845991] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.845997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102064 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.846004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.846012] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.846018] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.846024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102072 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.846031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.846038] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.846044] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.846050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102080 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.846057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.846065] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.846070] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.846076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102088 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.846086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.846093] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.846098] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.846105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:102096 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.846113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.846120] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.325 [2024-11-06 09:04:21.846125] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.325 [2024-11-06 09:04:21.846131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:101088 len:8 PRP1 0x0 PRP2 0x0 00:28:46.325 [2024-11-06 09:04:21.846138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:21.846181] bdev_nvme.c:2035:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:28:46.325 [2024-11-06 09:04:21.846191] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] in failed state. 00:28:46.325 [2024-11-06 09:04:21.846246] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x17f1d70 (9): Bad file descriptor 00:28:46.325 [2024-11-06 09:04:21.849792] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:28:46.325 [2024-11-06 09:04:21.982277] bdev_nvme.c:2236:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 3] Resetting controller successful. 00:28:46.325 10946.50 IOPS, 42.76 MiB/s [2024-11-06T08:04:36.438Z] 11165.33 IOPS, 43.61 MiB/s [2024-11-06T08:04:36.438Z] 11191.25 IOPS, 43.72 MiB/s [2024-11-06T08:04:36.438Z] [2024-11-06 09:04:25.339058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:54688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.325 [2024-11-06 09:04:25.339095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:25.339111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:55368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.325 [2024-11-06 09:04:25.339120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:25.339131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:55376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.325 [2024-11-06 09:04:25.339139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.325 [2024-11-06 09:04:25.339148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:55384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.325 [2024-11-06 09:04:25.339156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:55392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.326 [2024-11-06 09:04:25.339173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:55400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.326 [2024-11-06 09:04:25.339190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339199] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:55408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.326 [2024-11-06 09:04:25.339211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:55416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.326 [2024-11-06 09:04:25.339228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:55424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.326 [2024-11-06 09:04:25.339245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:55432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.326 [2024-11-06 09:04:25.339261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:55440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.326 [2024-11-06 09:04:25.339278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:55448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.326 [2024-11-06 09:04:25.339295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:55456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.326 [2024-11-06 09:04:25.339312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:55464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.326 [2024-11-06 09:04:25.339328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:54696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339354] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:54704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:54712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:54720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339394] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:54728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:54736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:54744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:54752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:54760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:54768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:54776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:54784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:54792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:54800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:54808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:54816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:54824 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:54832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:54840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:54848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:54856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339692] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:54864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:54872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:54880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:54888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:54896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:54904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339797] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:54912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:54920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:54928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.326 [2024-11-06 09:04:25.339838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.326 [2024-11-06 09:04:25.339848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:54936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.339857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.339867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:54944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.339874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.339883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:54952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.339891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.339900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:54960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.339908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.339917] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:54968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.339924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.339933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:54976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.339940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.339950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:54984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.339957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.339966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:54992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.339973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.339983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:55000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.339990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.339999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:55008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:55016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:55024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:55032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:55040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:55048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:55056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:55064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:55072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:55080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:55088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:55096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:55104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:55112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:55120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:55128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:55136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:55144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:55152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:55160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:55168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:55176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:55184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:55192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:55200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:55208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.327 [2024-11-06 09:04:25.340438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.327 [2024-11-06 09:04:25.340448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:55216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:55224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:55232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:55240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:55248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:55256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:55264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:55272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340576] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:55280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:55288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:55296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.328 [2024-11-06 09:04:25.340625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:55472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:55480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:55488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:55496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:55504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:55512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:55520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:55528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:55536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:55544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:55552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:55560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:55568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:55576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340871] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:55584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:55592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:55600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:55608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:55616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:55624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:55632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.340988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:55640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.340995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.341004] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:55648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.341011] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.341020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:55656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.341028] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.341037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:55664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.341044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.341053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:55672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.341061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.341070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:55680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.341077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.341087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:55688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.328 [2024-11-06 09:04:25.341095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.341117] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.328 [2024-11-06 09:04:25.341125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:55696 len:8 PRP1 0x0 PRP2 0x0 00:28:46.328 [2024-11-06 09:04:25.341133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.328 [2024-11-06 09:04:25.341143] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.329 [2024-11-06 09:04:25.341148] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.329 [2024-11-06 09:04:25.341157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:55704 len:8 PRP1 0x0 PRP2 0x0 00:28:46.329 [2024-11-06 09:04:25.341164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341172] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.329 [2024-11-06 09:04:25.341177] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.329 [2024-11-06 09:04:25.341183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:55304 len:8 PRP1 0x0 PRP2 0x0 00:28:46.329 [2024-11-06 09:04:25.341191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341198] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.329 [2024-11-06 09:04:25.341203] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.329 [2024-11-06 09:04:25.341210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:55312 len:8 PRP1 0x0 PRP2 0x0 00:28:46.329 [2024-11-06 09:04:25.341217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341224] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.329 [2024-11-06 09:04:25.341229] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.329 [2024-11-06 09:04:25.341235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:55320 len:8 PRP1 0x0 PRP2 0x0 00:28:46.329 [2024-11-06 09:04:25.341242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341249] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.329 [2024-11-06 09:04:25.341254] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.329 [2024-11-06 09:04:25.341261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:55328 len:8 PRP1 0x0 PRP2 0x0 00:28:46.329 [2024-11-06 09:04:25.341268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341276] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.329 [2024-11-06 09:04:25.341282] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.329 [2024-11-06 09:04:25.341288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:55336 len:8 PRP1 0x0 PRP2 0x0 00:28:46.329 [2024-11-06 09:04:25.341295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341303] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.329 [2024-11-06 09:04:25.341308] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.329 [2024-11-06 09:04:25.341315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:55344 len:8 PRP1 0x0 PRP2 0x0 00:28:46.329 [2024-11-06 09:04:25.341322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341329] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.329 [2024-11-06 09:04:25.341335] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.329 [2024-11-06 09:04:25.341341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:55352 len:8 PRP1 0x0 PRP2 0x0 00:28:46.329 [2024-11-06 09:04:25.341348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341355] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.329 [2024-11-06 09:04:25.341363] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.329 [2024-11-06 09:04:25.341369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:55360 len:8 PRP1 0x0 PRP2 0x0 00:28:46.329 [2024-11-06 09:04:25.341376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341413] bdev_nvme.c:2035:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 3] Start failover from 10.0.0.2:4421 to 10.0.0.2:4422 00:28:46.329 [2024-11-06 09:04:25.341434] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.329 [2024-11-06 09:04:25.341442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341450] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.329 [2024-11-06 09:04:25.341458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341466] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.329 [2024-11-06 09:04:25.341473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341482] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.329 [2024-11-06 09:04:25.341489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:25.341496] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 3] in failed state. 00:28:46.329 [2024-11-06 09:04:25.345059] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 3] resetting controller 00:28:46.329 [2024-11-06 09:04:25.345087] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x17f1d70 (9): Bad file descriptor 00:28:46.329 [2024-11-06 09:04:25.378453] bdev_nvme.c:2236:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 5] Resetting controller successful. 00:28:46.329 11151.80 IOPS, 43.56 MiB/s [2024-11-06T08:04:36.442Z] 11133.00 IOPS, 43.49 MiB/s [2024-11-06T08:04:36.442Z] 11128.43 IOPS, 43.47 MiB/s [2024-11-06T08:04:36.442Z] 11145.75 IOPS, 43.54 MiB/s [2024-11-06T08:04:36.442Z] [2024-11-06 09:04:29.719418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:56344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:56352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:56360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:56368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:56376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:56384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:56392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719579] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:56400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:56408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719603] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:56416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:56424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719637] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:56432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:56440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:56448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:57216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.329 [2024-11-06 09:04:29.719704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:56456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:56464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.329 [2024-11-06 09:04:29.719738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.329 [2024-11-06 09:04:29.719755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:56472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:56480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:56488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:56496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:56504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:56512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:56520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:56528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:56536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:56544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:56552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:56560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:56568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:56576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.719992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:56584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.719999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:56592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:56600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:56608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:56616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:56624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720094] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:56632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:57224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.330 [2024-11-06 09:04:29.720118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:56640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:56648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:56656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:56664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:56672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:56680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720225] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:56688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:56696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:56704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:56712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:56720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720309] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:56728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:56736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:56744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:56752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:56760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:56768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:56776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.330 [2024-11-06 09:04:29.720428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.330 [2024-11-06 09:04:29.720438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:56784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:56792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:56800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:56808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:56816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:56824 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:56832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:56840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:56848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:56856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:56864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:56872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:56880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:56888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:56896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720692] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:56904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:56912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:56920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:56928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:56936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:56944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720786] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:56952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:56960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:56968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:56976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720864] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:56984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:56992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:57000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720915] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:57008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720923] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:57016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:57024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:57032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.720984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:57040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.720991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.721001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:57048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.721008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.721018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:57056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.721026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.721036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:57064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.331 [2024-11-06 09:04:29.721043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.331 [2024-11-06 09:04:29.721054] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:57072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:57080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:57088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:57096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:57104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:57112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:57120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:57128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:57136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:57144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:57152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:57160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:57168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:57176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:57184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:57192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:57200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:57208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:28:46.332 [2024-11-06 09:04:29.721348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:57232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:57240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:57248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:57256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:57264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:57272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:57280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:57288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:57296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:57304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:57312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:57320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:57328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:57336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:57344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:57352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:28:46.332 [2024-11-06 09:04:29.721614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721636] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:28:46.332 [2024-11-06 09:04:29.721642] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:28:46.332 [2024-11-06 09:04:29.721649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:57360 len:8 PRP1 0x0 PRP2 0x0 00:28:46.332 [2024-11-06 09:04:29.721657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721701] bdev_nvme.c:2035:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 5] Start failover from 10.0.0.2:4422 to 10.0.0.2:4420 00:28:46.332 [2024-11-06 09:04:29.721723] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.332 [2024-11-06 09:04:29.721731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721740] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.332 [2024-11-06 09:04:29.721753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721762] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.332 [2024-11-06 09:04:29.721769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721777] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:28:46.332 [2024-11-06 09:04:29.721785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:46.332 [2024-11-06 09:04:29.721793] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 5] in failed state. 00:28:46.333 [2024-11-06 09:04:29.725365] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 5] resetting controller 00:28:46.333 [2024-11-06 09:04:29.725391] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x17f1d70 (9): Bad file descriptor 00:28:46.333 11065.44 IOPS, 43.22 MiB/s [2024-11-06T08:04:36.446Z] [2024-11-06 09:04:29.848426] bdev_nvme.c:2236:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 6] Resetting controller successful. 00:28:46.333 11049.30 IOPS, 43.16 MiB/s [2024-11-06T08:04:36.446Z] 11077.64 IOPS, 43.27 MiB/s [2024-11-06T08:04:36.446Z] 11086.25 IOPS, 43.31 MiB/s [2024-11-06T08:04:36.446Z] 11082.92 IOPS, 43.29 MiB/s [2024-11-06T08:04:36.446Z] 11114.86 IOPS, 43.42 MiB/s 00:28:46.333 Latency(us) 00:28:46.333 [2024-11-06T08:04:36.446Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:46.333 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:28:46.333 Verification LBA range: start 0x0 length 0x4000 00:28:46.333 NVMe0n1 : 15.01 11110.85 43.40 669.90 0.00 10837.17 570.03 29272.75 00:28:46.333 [2024-11-06T08:04:36.446Z] =================================================================================================================== 00:28:46.333 [2024-11-06T08:04:36.446Z] Total : 11110.85 43.40 669.90 0.00 10837.17 570.03 29272.75 00:28:46.333 Received shutdown signal, test time was about 15.000000 seconds 00:28:46.333 00:28:46.333 Latency(us) 00:28:46.333 [2024-11-06T08:04:36.446Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:46.333 [2024-11-06T08:04:36.446Z] =================================================================================================================== 00:28:46.333 [2024-11-06T08:04:36.446Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@65 -- # grep -c 'Resetting controller successful' 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@65 -- # count=3 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@67 -- # (( count != 3 )) 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@73 -- # bdevperf_pid=2006524 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@75 -- # waitforlisten 2006524 /var/tmp/bdevperf.sock 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 1 -f 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@831 -- # '[' -z 2006524 ']' 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@836 -- # local max_retries=100 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:28:46.333 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@840 -- # xtrace_disable 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@864 -- # return 0 00:28:46.333 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:28:46.333 [2024-11-06 09:04:36.406423] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:28:46.594 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:28:46.594 [2024-11-06 09:04:36.590883] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:28:46.594 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:28:46.855 NVMe0n1 00:28:46.855 09:04:36 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:28:47.116 00:28:47.377 09:04:37 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x failover 00:28:47.638 00:28:47.638 09:04:37 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:28:47.638 09:04:37 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@82 -- # grep -q NVMe0 00:28:47.638 09:04:37 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:28:47.899 09:04:37 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@87 -- # sleep 3 00:28:51.197 09:04:40 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:28:51.197 09:04:40 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@88 -- # grep -q NVMe0 00:28:51.197 09:04:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:28:51.197 09:04:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@90 -- # run_test_pid=2007441 00:28:51.197 09:04:41 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@92 -- # wait 2007441 00:28:52.139 { 00:28:52.139 "results": [ 00:28:52.139 { 00:28:52.139 "job": "NVMe0n1", 00:28:52.139 "core_mask": "0x1", 00:28:52.139 "workload": "verify", 00:28:52.139 "status": "finished", 00:28:52.139 "verify_range": { 00:28:52.139 "start": 0, 00:28:52.139 "length": 16384 00:28:52.139 }, 00:28:52.139 "queue_depth": 128, 00:28:52.139 "io_size": 4096, 00:28:52.139 "runtime": 1.010643, 00:28:52.139 "iops": 11569.86195916857, 00:28:52.139 "mibps": 45.19477327800222, 00:28:52.139 "io_failed": 0, 00:28:52.139 "io_timeout": 0, 00:28:52.139 "avg_latency_us": 10995.758647623934, 00:28:52.139 "min_latency_us": 1761.28, 00:28:52.139 "max_latency_us": 10212.693333333333 00:28:52.139 } 00:28:52.139 ], 00:28:52.139 "core_count": 1 00:28:52.139 } 00:28:52.139 09:04:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@94 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:28:52.139 [2024-11-06 09:04:36.059802] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:28:52.139 [2024-11-06 09:04:36.059862] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2006524 ] 00:28:52.139 [2024-11-06 09:04:36.130642] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:28:52.139 [2024-11-06 09:04:36.165321] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:28:52.139 [2024-11-06 09:04:37.860181] bdev_nvme.c:2035:bdev_nvme_failover_trid: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 7] Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:28:52.139 [2024-11-06 09:04:37.860227] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:28:52.139 [2024-11-06 09:04:37.860239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:52.139 [2024-11-06 09:04:37.860248] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:28:52.139 [2024-11-06 09:04:37.860256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:52.139 [2024-11-06 09:04:37.860264] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:28:52.139 [2024-11-06 09:04:37.860271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:52.139 [2024-11-06 09:04:37.860279] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:28:52.139 [2024-11-06 09:04:37.860287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:28:52.139 [2024-11-06 09:04:37.860294] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 7] in failed state. 00:28:52.139 [2024-11-06 09:04:37.860325] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 7] resetting controller 00:28:52.139 [2024-11-06 09:04:37.860340] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x7edd70 (9): Bad file descriptor 00:28:52.140 [2024-11-06 09:04:37.871258] bdev_nvme.c:2236:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 10] Resetting controller successful. 00:28:52.140 Running I/O for 1 seconds... 00:28:52.140 11529.00 IOPS, 45.04 MiB/s 00:28:52.140 Latency(us) 00:28:52.140 [2024-11-06T08:04:42.253Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:28:52.140 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:28:52.140 Verification LBA range: start 0x0 length 0x4000 00:28:52.140 NVMe0n1 : 1.01 11569.86 45.19 0.00 0.00 10995.76 1761.28 10212.69 00:28:52.140 [2024-11-06T08:04:42.253Z] =================================================================================================================== 00:28:52.140 [2024-11-06T08:04:42.253Z] Total : 11569.86 45.19 0.00 0.00 10995.76 1761.28 10212.69 00:28:52.140 09:04:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:28:52.140 09:04:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@95 -- # grep -q NVMe0 00:28:52.401 09:04:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@98 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:28:52.663 09:04:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:28:52.663 09:04:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@99 -- # grep -q NVMe0 00:28:52.663 09:04:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:28:52.923 09:04:42 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@101 -- # sleep 3 00:28:56.310 09:04:45 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:28:56.310 09:04:45 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@103 -- # grep -q NVMe0 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@108 -- # killprocess 2006524 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@950 -- # '[' -z 2006524 ']' 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # kill -0 2006524 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # uname 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2006524 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2006524' 00:28:56.310 killing process with pid 2006524 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@969 -- # kill 2006524 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@974 -- # wait 2006524 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@110 -- # sync 00:28:56.310 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@113 -- # trap - SIGINT SIGTERM EXIT 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@115 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- host/failover.sh@116 -- # nvmftestfini 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@514 -- # nvmfcleanup 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@121 -- # sync 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@124 -- # set +e 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@125 -- # for i in {1..20} 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:28:56.571 rmmod nvme_tcp 00:28:56.571 rmmod nvme_fabrics 00:28:56.571 rmmod nvme_keyring 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@128 -- # set -e 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@129 -- # return 0 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@515 -- # '[' -n 2002865 ']' 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@516 -- # killprocess 2002865 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@950 -- # '[' -z 2002865 ']' 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@954 -- # kill -0 2002865 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # uname 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2002865 00:28:56.571 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:28:56.572 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:28:56.572 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2002865' 00:28:56.572 killing process with pid 2002865 00:28:56.572 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@969 -- # kill 2002865 00:28:56.572 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@974 -- # wait 2002865 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@297 -- # iptr 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@789 -- # iptables-save 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@789 -- # iptables-restore 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@302 -- # remove_spdk_ns 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:56.833 09:04:46 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:58.748 09:04:48 nvmf_tcp.nvmf_host.nvmf_failover -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:28:58.748 00:28:58.748 real 0m39.237s 00:28:58.748 user 2m0.974s 00:28:58.748 sys 0m8.108s 00:28:58.748 09:04:48 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@1126 -- # xtrace_disable 00:28:58.748 09:04:48 nvmf_tcp.nvmf_host.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:28:58.748 ************************************ 00:28:58.748 END TEST nvmf_failover 00:28:58.748 ************************************ 00:28:59.010 09:04:48 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@26 -- # run_test nvmf_host_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:28:59.010 09:04:48 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:28:59.010 09:04:48 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:28:59.010 09:04:48 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:28:59.010 ************************************ 00:28:59.010 START TEST nvmf_host_discovery 00:28:59.010 ************************************ 00:28:59.010 09:04:48 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:28:59.010 * Looking for test storage... 00:28:59.010 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1689 -- # lcov --version 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@333 -- # local ver1 ver1_l 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@334 -- # local ver2 ver2_l 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@336 -- # IFS=.-: 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@336 -- # read -ra ver1 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@337 -- # IFS=.-: 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@337 -- # read -ra ver2 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@338 -- # local 'op=<' 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@340 -- # ver1_l=2 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@341 -- # ver2_l=1 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@344 -- # case "$op" in 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@345 -- # : 1 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@364 -- # (( v = 0 )) 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@365 -- # decimal 1 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@353 -- # local d=1 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@355 -- # echo 1 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@365 -- # ver1[v]=1 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@366 -- # decimal 2 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@353 -- # local d=2 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@355 -- # echo 2 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@366 -- # ver2[v]=2 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@368 -- # return 0 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:28:59.010 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:59.010 --rc genhtml_branch_coverage=1 00:28:59.010 --rc genhtml_function_coverage=1 00:28:59.010 --rc genhtml_legend=1 00:28:59.010 --rc geninfo_all_blocks=1 00:28:59.010 --rc geninfo_unexecuted_blocks=1 00:28:59.010 00:28:59.010 ' 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:28:59.010 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:59.010 --rc genhtml_branch_coverage=1 00:28:59.010 --rc genhtml_function_coverage=1 00:28:59.010 --rc genhtml_legend=1 00:28:59.010 --rc geninfo_all_blocks=1 00:28:59.010 --rc geninfo_unexecuted_blocks=1 00:28:59.010 00:28:59.010 ' 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:28:59.010 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:59.010 --rc genhtml_branch_coverage=1 00:28:59.010 --rc genhtml_function_coverage=1 00:28:59.010 --rc genhtml_legend=1 00:28:59.010 --rc geninfo_all_blocks=1 00:28:59.010 --rc geninfo_unexecuted_blocks=1 00:28:59.010 00:28:59.010 ' 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:28:59.010 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:28:59.010 --rc genhtml_branch_coverage=1 00:28:59.010 --rc genhtml_function_coverage=1 00:28:59.010 --rc genhtml_legend=1 00:28:59.010 --rc geninfo_all_blocks=1 00:28:59.010 --rc geninfo_unexecuted_blocks=1 00:28:59.010 00:28:59.010 ' 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@7 -- # uname -s 00:28:59.010 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@15 -- # shopt -s extglob 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@5 -- # export PATH 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@51 -- # : 0 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:28:59.272 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@55 -- # have_pci_nics=0 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@11 -- # '[' tcp == rdma ']' 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@16 -- # DISCOVERY_PORT=8009 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@17 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@20 -- # NQN=nqn.2016-06.io.spdk:cnode 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@22 -- # HOST_NQN=nqn.2021-12.io.spdk:test 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@23 -- # HOST_SOCK=/tmp/host.sock 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@25 -- # nvmftestinit 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@474 -- # prepare_net_devs 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@436 -- # local -g is_hw=no 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@438 -- # remove_spdk_ns 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@309 -- # xtrace_disable 00:28:59.272 09:04:49 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@315 -- # pci_devs=() 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@315 -- # local -a pci_devs 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@316 -- # pci_net_devs=() 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@317 -- # pci_drivers=() 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@317 -- # local -A pci_drivers 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@319 -- # net_devs=() 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@319 -- # local -ga net_devs 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@320 -- # e810=() 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@320 -- # local -ga e810 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@321 -- # x722=() 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@321 -- # local -ga x722 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@322 -- # mlx=() 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@322 -- # local -ga mlx 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:29:07.418 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:29:07.418 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@416 -- # [[ up == up ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:29:07.418 Found net devices under 0000:4b:00.0: cvl_0_0 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@416 -- # [[ up == up ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:29:07.418 Found net devices under 0000:4b:00.1: cvl_0_1 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@440 -- # is_hw=yes 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:29:07.418 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:29:07.419 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:29:07.419 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.626 ms 00:29:07.419 00:29:07.419 --- 10.0.0.2 ping statistics --- 00:29:07.419 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:07.419 rtt min/avg/max/mdev = 0.626/0.626/0.626/0.000 ms 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:29:07.419 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:29:07.419 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.274 ms 00:29:07.419 00:29:07.419 --- 10.0.0.1 ping statistics --- 00:29:07.419 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:07.419 rtt min/avg/max/mdev = 0.274/0.274/0.274/0.000 ms 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@448 -- # return 0 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@30 -- # nvmfappstart -m 0x2 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@724 -- # xtrace_disable 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@507 -- # nvmfpid=2012762 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@508 -- # waitforlisten 2012762 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@831 -- # '[' -z 2012762 ']' 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@836 -- # local max_retries=100 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:07.419 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@840 -- # xtrace_disable 00:29:07.419 09:04:56 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:07.419 [2024-11-06 09:04:56.616739] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:29:07.419 [2024-11-06 09:04:56.616813] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:07.419 [2024-11-06 09:04:56.716181] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:07.419 [2024-11-06 09:04:56.766319] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:29:07.419 [2024-11-06 09:04:56.766369] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:29:07.419 [2024-11-06 09:04:56.766378] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:29:07.419 [2024-11-06 09:04:56.766385] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:29:07.419 [2024-11-06 09:04:56.766392] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:29:07.419 [2024-11-06 09:04:56.767185] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@864 -- # return 0 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@730 -- # xtrace_disable 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@32 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:07.419 [2024-11-06 09:04:57.487492] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2014-08.org.nvmexpress.discovery -t tcp -a 10.0.0.2 -s 8009 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:07.419 [2024-11-06 09:04:57.499782] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@35 -- # rpc_cmd bdev_null_create null0 1000 512 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:07.419 null0 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@36 -- # rpc_cmd bdev_null_create null1 1000 512 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:07.419 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:07.419 null1 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@37 -- # rpc_cmd bdev_wait_for_examine 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@45 -- # hostpid=2012996 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@46 -- # waitforlisten 2012996 /tmp/host.sock 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@831 -- # '[' -z 2012996 ']' 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@835 -- # local rpc_addr=/tmp/host.sock 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@836 -- # local max_retries=100 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:29:07.680 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@840 -- # xtrace_disable 00:29:07.680 09:04:57 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:07.680 [2024-11-06 09:04:57.598639] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:29:07.680 [2024-11-06 09:04:57.598708] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2012996 ] 00:29:07.680 [2024-11-06 09:04:57.674343] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:07.680 [2024-11-06 09:04:57.716610] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@864 -- # return 0 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@48 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@50 -- # rpc_cmd -s /tmp/host.sock log_set_flag bdev_nvme 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@51 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@72 -- # notify_id=0 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@83 -- # get_subsystem_names 00:29:08.621 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@83 -- # [[ '' == '' ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@84 -- # get_bdev_list 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@84 -- # [[ '' == '' ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@86 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@87 -- # get_subsystem_names 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@87 -- # [[ '' == '' ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@88 -- # get_bdev_list 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@88 -- # [[ '' == '' ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@90 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@91 -- # get_subsystem_names 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@91 -- # [[ '' == '' ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@92 -- # get_bdev_list 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@92 -- # [[ '' == '' ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@96 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.622 [2024-11-06 09:04:58.718759] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@97 -- # get_subsystem_names 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.622 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:29:08.883 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.883 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@97 -- # [[ '' == '' ]] 00:29:08.883 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@98 -- # get_bdev_list 00:29:08.883 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:08.883 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:08.883 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@98 -- # [[ '' == '' ]] 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@99 -- # is_notification_count_eq 0 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=0 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@103 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2021-12.io.spdk:test 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@105 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_names 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ '' == \n\v\m\e\0 ]] 00:29:08.884 09:04:58 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # sleep 1 00:29:09.455 [2024-11-06 09:04:59.456144] bdev_nvme.c:7291:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:29:09.455 [2024-11-06 09:04:59.456164] bdev_nvme.c:7377:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:29:09.455 [2024-11-06 09:04:59.456178] bdev_nvme.c:7254:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:29:09.716 [2024-11-06 09:04:59.583581] bdev_nvme.c:7220:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:29:09.716 [2024-11-06 09:04:59.809025] bdev_nvme.c:5582:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr was created to 10.0.0.2:4420 00:29:09.716 [2024-11-06 09:04:59.809964] bdev_nvme.c:1963:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Connecting qpair 0x10ca750:1 started. 00:29:09.716 [2024-11-06 09:04:59.811574] bdev_nvme.c:7110:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:29:09.716 [2024-11-06 09:04:59.811599] bdev_nvme.c:7069:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:29:09.716 [2024-11-06 09:04:59.815427] bdev_nvme.c:1779:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpair 0x10ca750 was disconnected and freed. delete nvme_qpair. 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_names 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@106 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1"' ']]' 00:29:09.978 09:04:59 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_bdev_list 00:29:09.978 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:09.978 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:09.978 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:09.978 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:09.978 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0n1 == \n\v\m\e\0\n\1 ]] 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@107 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT"' ']]' 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_paths nvme0 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ 4420 == \4\4\2\0 ]] 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@108 -- # is_notification_count_eq 1 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=1 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:29:09.979 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=1 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=1 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@111 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null1 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@113 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_bdev_list 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:10.239 [2024-11-06 09:05:00.133436] bdev_nvme.c:1963:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Connecting qpair 0x1099030:1 started. 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:10.239 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:10.239 [2024-11-06 09:05:00.136637] bdev_nvme.c:1779:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpair 0x1099030 was disconnected and freed. delete nvme_qpair. 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@114 -- # is_notification_count_eq 1 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=1 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 1 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=1 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@118 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:10.240 [2024-11-06 09:05:00.234927] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:29:10.240 [2024-11-06 09:05:00.235660] bdev_nvme.c:7273:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:29:10.240 [2024-11-06 09:05:00.235682] bdev_nvme.c:7254:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@120 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_names 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@121 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_bdev_list 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:10.240 [2024-11-06 09:05:00.323941] bdev_nvme.c:7215:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new path for nvme0 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@122 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:10.240 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:29:10.500 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_paths nvme0 00:29:10.500 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:29:10.500 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:10.500 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:10.500 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:29:10.500 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:29:10.500 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:29:10.500 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:10.500 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ 4420 == \4\4\2\0\ \4\4\2\1 ]] 00:29:10.500 09:05:00 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@920 -- # sleep 1 00:29:10.500 [2024-11-06 09:05:00.423883] bdev_nvme.c:5582:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] ctrlr was created to 10.0.0.2:4421 00:29:10.500 [2024-11-06 09:05:00.423920] bdev_nvme.c:7110:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:29:10.500 [2024-11-06 09:05:00.423928] bdev_nvme.c:7069:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:29:10.500 [2024-11-06 09:05:00.423933] bdev_nvme.c:7069:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_paths nvme0 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ 4420 4421 == \4\4\2\0\ \4\4\2\1 ]] 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@123 -- # is_notification_count_eq 0 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@127 -- # rpc_cmd nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.443 [2024-11-06 09:05:01.494878] bdev_nvme.c:7273:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:29:11.443 [2024-11-06 09:05:01.494901] bdev_nvme.c:7254:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@129 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:29:11.443 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_names 00:29:11.443 [2024-11-06 09:05:01.503156] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:29:11.443 [2024-11-06 09:05:01.503176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:29:11.443 [2024-11-06 09:05:01.503186] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:29:11.443 [2024-11-06 09:05:01.503193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:29:11.444 [2024-11-06 09:05:01.503202] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:29:11.444 [2024-11-06 09:05:01.503209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:29:11.444 [2024-11-06 09:05:01.503217] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:29:11.444 [2024-11-06 09:05:01.503224] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:29:11.444 [2024-11-06 09:05:01.503232] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x109ae10 is same with the state(6) to be set 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.444 [2024-11-06 09:05:01.513168] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x109ae10 (9): Bad file descriptor 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.444 [2024-11-06 09:05:01.523206] bdev_nvme.c:2536:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:29:11.444 [2024-11-06 09:05:01.523218] bdev_nvme.c:2524:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:29:11.444 [2024-11-06 09:05:01.523224] bdev_nvme.c:2112:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:29:11.444 [2024-11-06 09:05:01.523229] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:29:11.444 [2024-11-06 09:05:01.523247] bdev_nvme.c:2508:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:29:11.444 [2024-11-06 09:05:01.523571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:29:11.444 [2024-11-06 09:05:01.523585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x109ae10 with addr=10.0.0.2, port=4420 00:29:11.444 [2024-11-06 09:05:01.523593] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x109ae10 is same with the state(6) to be set 00:29:11.444 [2024-11-06 09:05:01.523606] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x109ae10 (9): Bad file descriptor 00:29:11.444 [2024-11-06 09:05:01.523617] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:29:11.444 [2024-11-06 09:05:01.523624] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:29:11.444 [2024-11-06 09:05:01.523632] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:29:11.444 [2024-11-06 09:05:01.523639] bdev_nvme.c:2498:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:29:11.444 [2024-11-06 09:05:01.523649] bdev_nvme.c:2305:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:29:11.444 [2024-11-06 09:05:01.523659] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:29:11.444 [2024-11-06 09:05:01.533278] bdev_nvme.c:2536:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:29:11.444 [2024-11-06 09:05:01.533289] bdev_nvme.c:2524:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:29:11.444 [2024-11-06 09:05:01.533294] bdev_nvme.c:2112:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:29:11.444 [2024-11-06 09:05:01.533299] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:29:11.444 [2024-11-06 09:05:01.533313] bdev_nvme.c:2508:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:29:11.444 [2024-11-06 09:05:01.533619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:29:11.444 [2024-11-06 09:05:01.533631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x109ae10 with addr=10.0.0.2, port=4420 00:29:11.444 [2024-11-06 09:05:01.533638] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x109ae10 is same with the state(6) to be set 00:29:11.444 [2024-11-06 09:05:01.533649] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x109ae10 (9): Bad file descriptor 00:29:11.444 [2024-11-06 09:05:01.533660] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:29:11.444 [2024-11-06 09:05:01.533666] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:29:11.444 [2024-11-06 09:05:01.533674] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:29:11.444 [2024-11-06 09:05:01.533680] bdev_nvme.c:2498:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:29:11.444 [2024-11-06 09:05:01.533685] bdev_nvme.c:2305:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:29:11.444 [2024-11-06 09:05:01.533694] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:29:11.444 [2024-11-06 09:05:01.543344] bdev_nvme.c:2536:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:29:11.444 [2024-11-06 09:05:01.543358] bdev_nvme.c:2524:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:29:11.444 [2024-11-06 09:05:01.543363] bdev_nvme.c:2112:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:29:11.444 [2024-11-06 09:05:01.543367] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:29:11.444 [2024-11-06 09:05:01.543382] bdev_nvme.c:2508:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:29:11.444 [2024-11-06 09:05:01.543701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:29:11.444 [2024-11-06 09:05:01.543714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x109ae10 with addr=10.0.0.2, port=4420 00:29:11.444 [2024-11-06 09:05:01.543721] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x109ae10 is same with the state(6) to be set 00:29:11.444 [2024-11-06 09:05:01.543733] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x109ae10 (9): Bad file descriptor 00:29:11.444 [2024-11-06 09:05:01.543744] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:29:11.444 [2024-11-06 09:05:01.543756] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:29:11.444 [2024-11-06 09:05:01.543764] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:29:11.444 [2024-11-06 09:05:01.543773] bdev_nvme.c:2498:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:29:11.444 [2024-11-06 09:05:01.543778] bdev_nvme.c:2305:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:29:11.444 [2024-11-06 09:05:01.543787] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@130 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:11.444 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:29:11.444 [2024-11-06 09:05:01.553413] bdev_nvme.c:2536:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:29:11.444 [2024-11-06 09:05:01.553426] bdev_nvme.c:2524:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:29:11.444 [2024-11-06 09:05:01.553431] bdev_nvme.c:2112:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:29:11.444 [2024-11-06 09:05:01.553436] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:29:11.444 [2024-11-06 09:05:01.553449] bdev_nvme.c:2508:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:29:11.444 [2024-11-06 09:05:01.553764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:29:11.444 [2024-11-06 09:05:01.553777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x109ae10 with addr=10.0.0.2, port=4420 00:29:11.444 [2024-11-06 09:05:01.553784] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x109ae10 is same with the state(6) to be set 00:29:11.444 [2024-11-06 09:05:01.553795] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x109ae10 (9): Bad file descriptor 00:29:11.444 [2024-11-06 09:05:01.553805] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:29:11.444 [2024-11-06 09:05:01.553812] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:29:11.444 [2024-11-06 09:05:01.553819] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:29:11.444 [2024-11-06 09:05:01.553825] bdev_nvme.c:2498:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:29:11.444 [2024-11-06 09:05:01.553830] bdev_nvme.c:2305:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:29:11.444 [2024-11-06 09:05:01.553839] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:29:11.705 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_bdev_list 00:29:11.705 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:11.705 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:11.705 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:11.705 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.705 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.705 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:11.705 [2024-11-06 09:05:01.563933] bdev_nvme.c:2536:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:29:11.705 [2024-11-06 09:05:01.563949] bdev_nvme.c:2524:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:29:11.705 [2024-11-06 09:05:01.563954] bdev_nvme.c:2112:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:29:11.705 [2024-11-06 09:05:01.563959] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:29:11.705 [2024-11-06 09:05:01.563974] bdev_nvme.c:2508:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:29:11.705 [2024-11-06 09:05:01.564278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:29:11.705 [2024-11-06 09:05:01.564290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x109ae10 with addr=10.0.0.2, port=4420 00:29:11.705 [2024-11-06 09:05:01.564298] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x109ae10 is same with the state(6) to be set 00:29:11.705 [2024-11-06 09:05:01.564310] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x109ae10 (9): Bad file descriptor 00:29:11.705 [2024-11-06 09:05:01.564327] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:29:11.705 [2024-11-06 09:05:01.564334] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:29:11.705 [2024-11-06 09:05:01.564341] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:29:11.706 [2024-11-06 09:05:01.564347] bdev_nvme.c:2498:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:29:11.706 [2024-11-06 09:05:01.564352] bdev_nvme.c:2305:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:29:11.706 [2024-11-06 09:05:01.564361] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:29:11.706 [2024-11-06 09:05:01.574006] bdev_nvme.c:2536:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:29:11.706 [2024-11-06 09:05:01.574018] bdev_nvme.c:2524:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:29:11.706 [2024-11-06 09:05:01.574023] bdev_nvme.c:2112:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:29:11.706 [2024-11-06 09:05:01.574027] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:29:11.706 [2024-11-06 09:05:01.574041] bdev_nvme.c:2508:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:29:11.706 [2024-11-06 09:05:01.574345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:29:11.706 [2024-11-06 09:05:01.574356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x109ae10 with addr=10.0.0.2, port=4420 00:29:11.706 [2024-11-06 09:05:01.574364] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x109ae10 is same with the state(6) to be set 00:29:11.706 [2024-11-06 09:05:01.574374] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x109ae10 (9): Bad file descriptor 00:29:11.706 [2024-11-06 09:05:01.574392] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:29:11.706 [2024-11-06 09:05:01.574399] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:29:11.706 [2024-11-06 09:05:01.574406] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:29:11.706 [2024-11-06 09:05:01.574413] bdev_nvme.c:2498:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:29:11.706 [2024-11-06 09:05:01.574417] bdev_nvme.c:2305:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:29:11.706 [2024-11-06 09:05:01.574430] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:29:11.706 [2024-11-06 09:05:01.581416] bdev_nvme.c:7078:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 not found 00:29:11.706 [2024-11-06 09:05:01.581434] bdev_nvme.c:7069:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@131 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_SECOND_PORT"' ']]' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_paths nvme0 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ 4421 == \4\4\2\1 ]] 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@132 -- # is_notification_count_eq 0 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@134 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_stop_discovery -b nvme 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@136 -- # waitforcondition '[[ "$(get_subsystem_names)" == "" ]]' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_subsystem_names)" == "" ]]' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_subsystem_names)"' == '""' ']]' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_subsystem_names 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:29:11.706 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ '' == '' ]] 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@137 -- # waitforcondition '[[ "$(get_bdev_list)" == "" ]]' 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=[[ "$(get_bdev_list)" == "" ]]' 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval '[[' '"$(get_bdev_list)"' == '""' ']]' 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_bdev_list 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:11.707 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # [[ '' == '' ]] 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@138 -- # is_notification_count_eq 2 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=2 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@914 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@915 -- # local max=10 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@916 -- # (( max-- )) 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # get_notification_count 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=2 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=4 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@917 -- # (( notification_count == expected_count )) 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@918 -- # return 0 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@141 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:11.967 09:05:01 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:12.909 [2024-11-06 09:05:02.936836] bdev_nvme.c:7291:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:29:12.910 [2024-11-06 09:05:02.936853] bdev_nvme.c:7377:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:29:12.910 [2024-11-06 09:05:02.936866] bdev_nvme.c:7254:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:29:13.170 [2024-11-06 09:05:03.025141] bdev_nvme.c:7220:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new subsystem nvme0 00:29:13.431 [2024-11-06 09:05:03.292506] bdev_nvme.c:5582:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 3] ctrlr was created to 10.0.0.2:4421 00:29:13.431 [2024-11-06 09:05:03.293143] bdev_nvme.c:1963:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 3] Connecting qpair 0x1097d20:1 started. 00:29:13.431 [2024-11-06 09:05:03.294967] bdev_nvme.c:7110:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:29:13.431 [2024-11-06 09:05:03.294994] bdev_nvme.c:7069:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:29:13.431 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.431 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@143 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:29:13.431 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@650 -- # local es=0 00:29:13.431 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:29:13.431 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:29:13.431 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:13.432 [2024-11-06 09:05:03.303730] bdev_nvme.c:1779:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 3] qpair 0x1097d20 was disconnected and freed. delete nvme_qpair. 00:29:13.432 request: 00:29:13.432 { 00:29:13.432 "name": "nvme", 00:29:13.432 "trtype": "tcp", 00:29:13.432 "traddr": "10.0.0.2", 00:29:13.432 "adrfam": "ipv4", 00:29:13.432 "trsvcid": "8009", 00:29:13.432 "hostnqn": "nqn.2021-12.io.spdk:test", 00:29:13.432 "wait_for_attach": true, 00:29:13.432 "method": "bdev_nvme_start_discovery", 00:29:13.432 "req_id": 1 00:29:13.432 } 00:29:13.432 Got JSON-RPC error response 00:29:13.432 response: 00:29:13.432 { 00:29:13.432 "code": -17, 00:29:13.432 "message": "File exists" 00:29:13.432 } 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # es=1 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@145 -- # get_discovery_ctrlrs 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@145 -- # [[ nvme == \n\v\m\e ]] 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@146 -- # get_bdev_list 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@146 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@149 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@650 -- # local es=0 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:13.432 request: 00:29:13.432 { 00:29:13.432 "name": "nvme_second", 00:29:13.432 "trtype": "tcp", 00:29:13.432 "traddr": "10.0.0.2", 00:29:13.432 "adrfam": "ipv4", 00:29:13.432 "trsvcid": "8009", 00:29:13.432 "hostnqn": "nqn.2021-12.io.spdk:test", 00:29:13.432 "wait_for_attach": true, 00:29:13.432 "method": "bdev_nvme_start_discovery", 00:29:13.432 "req_id": 1 00:29:13.432 } 00:29:13.432 Got JSON-RPC error response 00:29:13.432 response: 00:29:13.432 { 00:29:13.432 "code": -17, 00:29:13.432 "message": "File exists" 00:29:13.432 } 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # es=1 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@151 -- # get_discovery_ctrlrs 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@151 -- # [[ nvme == \n\v\m\e ]] 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@152 -- # get_bdev_list 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:29:13.432 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@152 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@155 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@650 -- # local es=0 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:13.693 09:05:03 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:14.634 [2024-11-06 09:05:04.558475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:29:14.634 [2024-11-06 09:05:04.558502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x10b0050 with addr=10.0.0.2, port=8010 00:29:14.634 [2024-11-06 09:05:04.558515] nvme_tcp.c:2612:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:29:14.634 [2024-11-06 09:05:04.558522] nvme.c: 831:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:29:14.634 [2024-11-06 09:05:04.558529] bdev_nvme.c:7359:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:29:15.575 [2024-11-06 09:05:05.560808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:29:15.575 [2024-11-06 09:05:05.560834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x10b0050 with addr=10.0.0.2, port=8010 00:29:15.575 [2024-11-06 09:05:05.560845] nvme_tcp.c:2612:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:29:15.575 [2024-11-06 09:05:05.560852] nvme.c: 831:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:29:15.575 [2024-11-06 09:05:05.560859] bdev_nvme.c:7359:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:29:16.516 [2024-11-06 09:05:06.562810] bdev_nvme.c:7334:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] timed out while attaching discovery ctrlr 00:29:16.516 request: 00:29:16.516 { 00:29:16.516 "name": "nvme_second", 00:29:16.516 "trtype": "tcp", 00:29:16.516 "traddr": "10.0.0.2", 00:29:16.516 "adrfam": "ipv4", 00:29:16.516 "trsvcid": "8010", 00:29:16.516 "hostnqn": "nqn.2021-12.io.spdk:test", 00:29:16.516 "wait_for_attach": false, 00:29:16.516 "attach_timeout_ms": 3000, 00:29:16.516 "method": "bdev_nvme_start_discovery", 00:29:16.516 "req_id": 1 00:29:16.516 } 00:29:16.516 Got JSON-RPC error response 00:29:16.516 response: 00:29:16.516 { 00:29:16.516 "code": -110, 00:29:16.516 "message": "Connection timed out" 00:29:16.516 } 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@653 -- # es=1 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@157 -- # get_discovery_ctrlrs 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@561 -- # xtrace_disable 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@157 -- # [[ nvme == \n\v\m\e ]] 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@159 -- # trap - SIGINT SIGTERM EXIT 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@161 -- # kill 2012996 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- host/discovery.sh@162 -- # nvmftestfini 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@514 -- # nvmfcleanup 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@121 -- # sync 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@124 -- # set +e 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@125 -- # for i in {1..20} 00:29:16.516 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:29:16.777 rmmod nvme_tcp 00:29:16.777 rmmod nvme_fabrics 00:29:16.778 rmmod nvme_keyring 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@128 -- # set -e 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@129 -- # return 0 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@515 -- # '[' -n 2012762 ']' 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@516 -- # killprocess 2012762 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@950 -- # '[' -z 2012762 ']' 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@954 -- # kill -0 2012762 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@955 -- # uname 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2012762 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2012762' 00:29:16.778 killing process with pid 2012762 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@969 -- # kill 2012762 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@974 -- # wait 2012762 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@297 -- # iptr 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@789 -- # iptables-restore 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@789 -- # iptables-save 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@302 -- # remove_spdk_ns 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:16.778 09:05:06 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:19.328 09:05:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:29:19.328 00:29:19.328 real 0m20.018s 00:29:19.328 user 0m23.226s 00:29:19.328 sys 0m6.995s 00:29:19.328 09:05:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@1126 -- # xtrace_disable 00:29:19.328 09:05:08 nvmf_tcp.nvmf_host.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:29:19.328 ************************************ 00:29:19.328 END TEST nvmf_host_discovery 00:29:19.328 ************************************ 00:29:19.328 09:05:08 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@27 -- # run_test nvmf_host_multipath_status /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multipath_status.sh --transport=tcp 00:29:19.328 09:05:08 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:29:19.328 09:05:08 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:29:19.328 09:05:08 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:29:19.328 ************************************ 00:29:19.328 START TEST nvmf_host_multipath_status 00:29:19.328 ************************************ 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multipath_status.sh --transport=tcp 00:29:19.328 * Looking for test storage... 00:29:19.328 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1689 -- # lcov --version 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@333 -- # local ver1 ver1_l 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@334 -- # local ver2 ver2_l 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@336 -- # IFS=.-: 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@336 -- # read -ra ver1 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@337 -- # IFS=.-: 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@337 -- # read -ra ver2 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@338 -- # local 'op=<' 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@340 -- # ver1_l=2 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@341 -- # ver2_l=1 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@344 -- # case "$op" in 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@345 -- # : 1 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@364 -- # (( v = 0 )) 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@365 -- # decimal 1 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@353 -- # local d=1 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@355 -- # echo 1 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@365 -- # ver1[v]=1 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@366 -- # decimal 2 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@353 -- # local d=2 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@355 -- # echo 2 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@366 -- # ver2[v]=2 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@368 -- # return 0 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:29:19.328 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:19.328 --rc genhtml_branch_coverage=1 00:29:19.328 --rc genhtml_function_coverage=1 00:29:19.328 --rc genhtml_legend=1 00:29:19.328 --rc geninfo_all_blocks=1 00:29:19.328 --rc geninfo_unexecuted_blocks=1 00:29:19.328 00:29:19.328 ' 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:29:19.328 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:19.328 --rc genhtml_branch_coverage=1 00:29:19.328 --rc genhtml_function_coverage=1 00:29:19.328 --rc genhtml_legend=1 00:29:19.328 --rc geninfo_all_blocks=1 00:29:19.328 --rc geninfo_unexecuted_blocks=1 00:29:19.328 00:29:19.328 ' 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:29:19.328 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:19.328 --rc genhtml_branch_coverage=1 00:29:19.328 --rc genhtml_function_coverage=1 00:29:19.328 --rc genhtml_legend=1 00:29:19.328 --rc geninfo_all_blocks=1 00:29:19.328 --rc geninfo_unexecuted_blocks=1 00:29:19.328 00:29:19.328 ' 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:29:19.328 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:19.328 --rc genhtml_branch_coverage=1 00:29:19.328 --rc genhtml_function_coverage=1 00:29:19.328 --rc genhtml_legend=1 00:29:19.328 --rc geninfo_all_blocks=1 00:29:19.328 --rc geninfo_unexecuted_blocks=1 00:29:19.328 00:29:19.328 ' 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@7 -- # uname -s 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:29:19.328 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@15 -- # shopt -s extglob 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@5 -- # export PATH 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@51 -- # : 0 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:29:19.329 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@55 -- # have_pci_nics=0 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@12 -- # MALLOC_BDEV_SIZE=64 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@16 -- # bpf_sh=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/bpftrace.sh 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@18 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@21 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@31 -- # nvmftestinit 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@474 -- # prepare_net_devs 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@436 -- # local -g is_hw=no 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@438 -- # remove_spdk_ns 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@309 -- # xtrace_disable 00:29:19.329 09:05:09 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@315 -- # pci_devs=() 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@315 -- # local -a pci_devs 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@316 -- # pci_net_devs=() 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@317 -- # pci_drivers=() 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@317 -- # local -A pci_drivers 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@319 -- # net_devs=() 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@319 -- # local -ga net_devs 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@320 -- # e810=() 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@320 -- # local -ga e810 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@321 -- # x722=() 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@321 -- # local -ga x722 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@322 -- # mlx=() 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@322 -- # local -ga mlx 00:29:27.481 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:29:27.482 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:29:27.482 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@416 -- # [[ up == up ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:29:27.482 Found net devices under 0000:4b:00.0: cvl_0_0 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@416 -- # [[ up == up ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:29:27.482 Found net devices under 0000:4b:00.1: cvl_0_1 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@440 -- # is_hw=yes 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:29:27.482 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:29:27.482 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.693 ms 00:29:27.482 00:29:27.482 --- 10.0.0.2 ping statistics --- 00:29:27.482 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:27.482 rtt min/avg/max/mdev = 0.693/0.693/0.693/0.000 ms 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:29:27.482 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:29:27.482 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.321 ms 00:29:27.482 00:29:27.482 --- 10.0.0.1 ping statistics --- 00:29:27.482 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:27.482 rtt min/avg/max/mdev = 0.321/0.321/0.321/0.000 ms 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@448 -- # return 0 00:29:27.482 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@33 -- # nvmfappstart -m 0x3 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@724 -- # xtrace_disable 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@507 -- # nvmfpid=2018986 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@508 -- # waitforlisten 2018986 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@831 -- # '[' -z 2018986 ']' 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@836 -- # local max_retries=100 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:27.483 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@840 -- # xtrace_disable 00:29:27.483 09:05:16 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:29:27.483 [2024-11-06 09:05:16.559235] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:29:27.483 [2024-11-06 09:05:16.559308] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:27.483 [2024-11-06 09:05:16.641833] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:29:27.483 [2024-11-06 09:05:16.683391] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:29:27.483 [2024-11-06 09:05:16.683427] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:29:27.483 [2024-11-06 09:05:16.683435] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:29:27.483 [2024-11-06 09:05:16.683442] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:29:27.483 [2024-11-06 09:05:16.683448] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:29:27.483 [2024-11-06 09:05:16.684807] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:29:27.483 [2024-11-06 09:05:16.685001] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:29:27.483 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:29:27.483 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@864 -- # return 0 00:29:27.483 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:29:27.483 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@730 -- # xtrace_disable 00:29:27.483 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:29:27.483 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:29:27.483 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@34 -- # nvmfapp_pid=2018986 00:29:27.483 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:29:27.483 [2024-11-06 09:05:17.540667] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:29:27.483 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:29:27.744 Malloc0 00:29:27.744 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -r -m 2 00:29:28.005 09:05:17 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:29:28.005 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:29:28.266 [2024-11-06 09:05:18.232398] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:29:28.266 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:29:28.528 [2024-11-06 09:05:18.388769] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 90 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@45 -- # bdevperf_pid=2019399 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@47 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@48 -- # waitforlisten 2019399 /var/tmp/bdevperf.sock 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@831 -- # '[' -z 2019399 ']' 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@836 -- # local max_retries=100 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:29:28.528 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@840 -- # xtrace_disable 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@864 -- # return 0 00:29:28.528 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_set_options -r -1 00:29:28.790 09:05:18 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x multipath -l -1 -o 10 00:29:29.363 Nvme0n1 00:29:29.363 09:05:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x multipath -l -1 -o 10 00:29:29.624 Nvme0n1 00:29:29.624 09:05:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 120 -s /var/tmp/bdevperf.sock perform_tests 00:29:29.624 09:05:19 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@78 -- # sleep 2 00:29:31.540 09:05:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@90 -- # set_ANA_state optimized optimized 00:29:31.540 09:05:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n optimized 00:29:31.804 09:05:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:29:32.105 09:05:21 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@91 -- # sleep 1 00:29:33.050 09:05:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@92 -- # check_status true false true true true true 00:29:33.050 09:05:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:29:33.050 09:05:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:33.050 09:05:22 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:29:33.312 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:33.312 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:29:33.312 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:29:33.312 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:33.312 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:33.312 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:29:33.312 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:33.312 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:29:33.573 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:33.573 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:29:33.573 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:33.573 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:29:33.835 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:33.835 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:29:33.835 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:33.835 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:29:33.835 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:33.835 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:29:33.835 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:33.835 09:05:23 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:29:34.097 09:05:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:34.097 09:05:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@94 -- # set_ANA_state non_optimized optimized 00:29:34.097 09:05:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:29:34.358 09:05:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:29:34.618 09:05:24 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@95 -- # sleep 1 00:29:35.560 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@96 -- # check_status false true true true true true 00:29:35.560 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:29:35.560 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:35.560 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:29:35.560 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:35.560 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:29:35.560 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:35.560 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:29:35.821 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:35.821 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:29:35.821 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:35.821 09:05:25 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:29:36.082 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:36.082 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:29:36.082 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:36.082 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:29:36.082 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:36.082 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:29:36.082 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:36.082 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:29:36.342 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:36.342 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:29:36.342 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:36.342 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:29:36.601 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:36.601 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@100 -- # set_ANA_state non_optimized non_optimized 00:29:36.601 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:29:36.861 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n non_optimized 00:29:36.861 09:05:26 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@101 -- # sleep 1 00:29:38.246 09:05:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@102 -- # check_status true false true true true true 00:29:38.246 09:05:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:29:38.246 09:05:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:38.246 09:05:27 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:29:38.246 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:38.246 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:29:38.246 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:38.247 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:29:38.247 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:38.247 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:29:38.247 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:38.247 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:29:38.508 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:38.508 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:29:38.508 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:38.508 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:29:38.769 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:38.769 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:29:38.769 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:38.769 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:29:38.769 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:38.769 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:29:38.769 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:38.769 09:05:28 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:29:39.031 09:05:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:39.031 09:05:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@104 -- # set_ANA_state non_optimized inaccessible 00:29:39.031 09:05:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:29:39.292 09:05:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:29:39.553 09:05:29 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@105 -- # sleep 1 00:29:40.494 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@106 -- # check_status true false true true true false 00:29:40.494 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:29:40.494 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:40.494 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:29:40.755 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:40.755 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:29:40.755 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:40.755 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:29:40.755 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:40.755 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:29:40.755 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:40.755 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:29:41.016 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:41.016 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:29:41.016 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:41.016 09:05:30 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:29:41.016 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:41.016 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:29:41.016 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:41.277 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:29:41.277 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:41.277 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:29:41.277 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:41.277 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:29:41.537 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:41.537 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@108 -- # set_ANA_state inaccessible inaccessible 00:29:41.537 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n inaccessible 00:29:41.798 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:29:41.798 09:05:31 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@109 -- # sleep 1 00:29:43.185 09:05:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@110 -- # check_status false false true true false false 00:29:43.185 09:05:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:29:43.185 09:05:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:43.185 09:05:32 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:29:43.185 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:43.185 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:29:43.185 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:43.185 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:29:43.185 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:43.185 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:29:43.185 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:43.185 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:29:43.446 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:43.446 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:29:43.446 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:43.446 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:29:43.707 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:43.707 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible false 00:29:43.707 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:43.707 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:29:43.707 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:43.707 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:29:43.707 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:43.707 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:29:43.968 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:43.968 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@112 -- # set_ANA_state inaccessible optimized 00:29:43.968 09:05:33 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n inaccessible 00:29:44.229 09:05:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:29:44.229 09:05:34 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@113 -- # sleep 1 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@114 -- # check_status false true true true false true 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:45.613 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:29:45.872 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:45.872 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:29:45.872 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:45.872 09:05:35 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:29:46.131 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:46.131 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible false 00:29:46.131 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:46.131 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:29:46.131 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:46.131 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:29:46.131 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:46.131 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:29:46.390 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:46.390 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@116 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_set_multipath_policy -b Nvme0n1 -p active_active 00:29:46.649 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@119 -- # set_ANA_state optimized optimized 00:29:46.650 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n optimized 00:29:46.909 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:29:46.909 09:05:36 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@120 -- # sleep 1 00:29:48.292 09:05:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@121 -- # check_status true true true true true true 00:29:48.292 09:05:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:29:48.292 09:05:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:48.292 09:05:37 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:29:48.292 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:48.292 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:29:48.292 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:48.292 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:29:48.292 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:48.292 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:29:48.292 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:29:48.292 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:48.553 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:48.553 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:29:48.553 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:29:48.553 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:48.813 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:48.813 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:29:48.813 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:48.813 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:29:48.813 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:48.813 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:29:48.813 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:48.813 09:05:38 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:29:49.073 09:05:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:49.073 09:05:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@123 -- # set_ANA_state non_optimized optimized 00:29:49.073 09:05:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:29:49.333 09:05:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:29:49.593 09:05:39 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@124 -- # sleep 1 00:29:50.532 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@125 -- # check_status false true true true true true 00:29:50.532 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:29:50.532 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:50.532 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:29:50.792 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:50.793 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:29:50.793 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:50.793 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:29:50.793 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:50.793 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:29:50.793 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:50.793 09:05:40 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:29:51.053 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:51.053 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:29:51.053 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:51.053 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:29:51.314 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:51.314 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:29:51.314 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:51.314 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:29:51.314 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:51.314 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:29:51.314 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:51.314 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:29:51.574 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:51.574 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@129 -- # set_ANA_state non_optimized non_optimized 00:29:51.574 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:29:51.834 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n non_optimized 00:29:52.094 09:05:41 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@130 -- # sleep 1 00:29:53.035 09:05:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@131 -- # check_status true true true true true true 00:29:53.035 09:05:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:29:53.035 09:05:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:29:53.035 09:05:42 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:53.294 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:53.294 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:29:53.294 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:53.294 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:29:53.294 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:53.294 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:29:53.294 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:53.294 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:29:53.555 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:53.555 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:29:53.555 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:53.555 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:29:53.815 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:53.815 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:29:53.815 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:53.815 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:29:53.815 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:53.815 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:29:53.815 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:53.815 09:05:43 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:29:54.075 09:05:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:54.075 09:05:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@133 -- # set_ANA_state non_optimized inaccessible 00:29:54.075 09:05:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:29:54.335 09:05:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:29:54.595 09:05:44 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@134 -- # sleep 1 00:29:55.534 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@135 -- # check_status true false true true true false 00:29:55.535 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:29:55.535 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:55.535 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:29:55.795 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:55.795 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:29:55.795 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:55.795 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:29:55.795 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:55.795 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:29:55.795 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:55.795 09:05:45 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:29:56.056 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:56.056 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:29:56.056 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:56.056 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:29:56.316 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:56.316 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:29:56.316 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:56.316 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:29:56.316 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:29:56.316 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:29:56.316 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:29:56.316 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@137 -- # killprocess 2019399 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@950 -- # '[' -z 2019399 ']' 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # kill -0 2019399 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@955 -- # uname 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2019399 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@956 -- # process_name=reactor_2 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # '[' reactor_2 = sudo ']' 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2019399' 00:29:56.577 killing process with pid 2019399 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@969 -- # kill 2019399 00:29:56.577 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@974 -- # wait 2019399 00:29:56.577 { 00:29:56.577 "results": [ 00:29:56.577 { 00:29:56.577 "job": "Nvme0n1", 00:29:56.577 "core_mask": "0x4", 00:29:56.577 "workload": "verify", 00:29:56.577 "status": "terminated", 00:29:56.577 "verify_range": { 00:29:56.577 "start": 0, 00:29:56.577 "length": 16384 00:29:56.577 }, 00:29:56.577 "queue_depth": 128, 00:29:56.577 "io_size": 4096, 00:29:56.577 "runtime": 26.91226, 00:29:56.577 "iops": 10759.2227482939, 00:29:56.577 "mibps": 42.028213860523046, 00:29:56.577 "io_failed": 0, 00:29:56.577 "io_timeout": 0, 00:29:56.577 "avg_latency_us": 11877.584091266484, 00:29:56.577 "min_latency_us": 226.98666666666668, 00:29:56.577 "max_latency_us": 3019898.88 00:29:56.577 } 00:29:56.577 ], 00:29:56.577 "core_count": 1 00:29:56.577 } 00:29:56.841 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@139 -- # wait 2019399 00:29:56.841 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@141 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:29:56.841 [2024-11-06 09:05:18.438188] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:29:56.841 [2024-11-06 09:05:18.438247] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2019399 ] 00:29:56.841 [2024-11-06 09:05:18.496384] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:56.841 [2024-11-06 09:05:18.525185] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:29:56.841 Running I/O for 90 seconds... 00:29:56.841 9443.00 IOPS, 36.89 MiB/s [2024-11-06T08:05:46.954Z] 9548.50 IOPS, 37.30 MiB/s [2024-11-06T08:05:46.954Z] 9602.00 IOPS, 37.51 MiB/s [2024-11-06T08:05:46.954Z] 9626.25 IOPS, 37.60 MiB/s [2024-11-06T08:05:46.954Z] 9883.20 IOPS, 38.61 MiB/s [2024-11-06T08:05:46.954Z] 10362.00 IOPS, 40.48 MiB/s [2024-11-06T08:05:46.954Z] 10731.29 IOPS, 41.92 MiB/s [2024-11-06T08:05:46.954Z] 10697.50 IOPS, 41.79 MiB/s [2024-11-06T08:05:46.954Z] 10578.44 IOPS, 41.32 MiB/s [2024-11-06T08:05:46.954Z] 10491.20 IOPS, 40.98 MiB/s [2024-11-06T08:05:46.954Z] 10416.27 IOPS, 40.69 MiB/s [2024-11-06T08:05:46.954Z] [2024-11-06 09:05:31.658143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:72096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.841 [2024-11-06 09:05:31.658179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:29:56.841 [2024-11-06 09:05:31.658213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:72104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:72112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:72120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:72128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:72136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:72144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:72152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:72160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:72168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658476] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:72176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:72184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:72192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:72200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:72208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:72216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:72224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:72232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:72240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658730] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:72248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:72256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:72264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658797] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:72272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.658813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:72280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.658818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:72288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.659222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:71592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.842 [2024-11-06 09:05:31.659240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:71600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.842 [2024-11-06 09:05:31.659257] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:71608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.842 [2024-11-06 09:05:31.659273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:71616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.842 [2024-11-06 09:05:31.659290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:71624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.842 [2024-11-06 09:05:31.659307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:71632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.842 [2024-11-06 09:05:31.659324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:71640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.842 [2024-11-06 09:05:31.659341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:72296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.659357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:72304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.659374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:72312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.659392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:72320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.659410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:72328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.659427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:29:56.842 [2024-11-06 09:05:31.659438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:72336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.842 [2024-11-06 09:05:31.659443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:72344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:72352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:72360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:72368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:72376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659612] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:72384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:72392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:72400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:72408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:72416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:72424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:72432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:72440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:72448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:72456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:72464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.659945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:72472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.659950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:72480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:72488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:72496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:72504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:72512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:72520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:72528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:72536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:72544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:72552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:72560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:72568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:72576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:72584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:72592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.660555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:72600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.660560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.662024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:72608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.843 [2024-11-06 09:05:31.662033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.662048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:71648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.843 [2024-11-06 09:05:31.662054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.662067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:71656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.843 [2024-11-06 09:05:31.662072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.662086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:71664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.843 [2024-11-06 09:05:31.662092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.662105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:71672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.843 [2024-11-06 09:05:31.662111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:29:56.843 [2024-11-06 09:05:31.662124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:71680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.843 [2024-11-06 09:05:31.662129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:71688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662162] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:71696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:71704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662199] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:71712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:71720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:71728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:71736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662277] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:71744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:71752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:71760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:71768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:71776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:71784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:71792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662445] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:71800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:71808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:71816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:71824 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:71832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:71840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:71848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:71856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:71864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:71872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662629] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:71880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:71888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:71896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:71904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:71912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:71920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:71928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:71936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:71944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:29:56.844 [2024-11-06 09:05:31.662825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:71952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.844 [2024-11-06 09:05:31.662831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.662846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:71960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.662851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.662865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:71968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.662870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.662885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:71976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.662890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.662905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:71984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.662910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.662925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:71992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.662930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.662944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:72000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.662950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.662964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:72008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.662969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.662984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:72016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.662990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.663004] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:72024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.663010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.663025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:72032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.663030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.663044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:72040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.663050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.663065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:72048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.663070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.663084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:72056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.663089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.663104] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:72064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.663109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.663124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:72072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.663129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.663143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:72080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.663149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:31.663163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:72088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:31.663168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:29:56.845 10303.08 IOPS, 40.25 MiB/s [2024-11-06T08:05:46.958Z] 9510.54 IOPS, 37.15 MiB/s [2024-11-06T08:05:46.958Z] 8831.21 IOPS, 34.50 MiB/s [2024-11-06T08:05:46.958Z] 8292.27 IOPS, 32.39 MiB/s [2024-11-06T08:05:46.958Z] 8587.69 IOPS, 33.55 MiB/s [2024-11-06T08:05:46.958Z] 8832.82 IOPS, 34.50 MiB/s [2024-11-06T08:05:46.958Z] 9257.44 IOPS, 36.16 MiB/s [2024-11-06T08:05:46.958Z] 9653.68 IOPS, 37.71 MiB/s [2024-11-06T08:05:46.958Z] 9934.25 IOPS, 38.81 MiB/s [2024-11-06T08:05:46.958Z] 10068.81 IOPS, 39.33 MiB/s [2024-11-06T08:05:46.958Z] 10194.09 IOPS, 39.82 MiB/s [2024-11-06T08:05:46.958Z] 10432.09 IOPS, 40.75 MiB/s [2024-11-06T08:05:46.958Z] 10697.00 IOPS, 41.79 MiB/s [2024-11-06T08:05:46.958Z] [2024-11-06 09:05:44.423330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:47576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.423371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.423403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:47592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.423410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.423420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:47608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.423431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.423442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:47624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.423447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.423457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:47640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.423462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.423473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:47656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.423478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.423488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:47672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.423493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.423504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:47688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.423509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.423520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:47560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.845 [2024-11-06 09:05:44.423525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:47712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.424599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:47728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.424617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:47744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.424633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424645] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:47760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.424651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:47776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.424666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:47792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.424682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:47808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.424701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:47824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.424716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:47840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.424732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:47856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.845 [2024-11-06 09:05:44.424752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:29:56.845 [2024-11-06 09:05:44.424763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:47872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.846 [2024-11-06 09:05:44.424768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:29:56.846 [2024-11-06 09:05:44.424779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:47888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.846 [2024-11-06 09:05:44.424784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:29:56.846 [2024-11-06 09:05:44.424794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:47904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.846 [2024-11-06 09:05:44.424800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:29:56.846 [2024-11-06 09:05:44.424810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:47920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.846 [2024-11-06 09:05:44.424815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:29:56.846 [2024-11-06 09:05:44.424826] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:47936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:29:56.846 [2024-11-06 09:05:44.424831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:29:56.846 [2024-11-06 09:05:44.424841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:47016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.846 [2024-11-06 09:05:44.424847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:29:56.846 [2024-11-06 09:05:44.424860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:47056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.846 [2024-11-06 09:05:44.424866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:29:56.846 [2024-11-06 09:05:44.424877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:47096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.846 [2024-11-06 09:05:44.424882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:29:56.846 [2024-11-06 09:05:44.424895] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:47128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:29:56.846 [2024-11-06 09:05:44.424900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:29:56.846 10847.04 IOPS, 42.37 MiB/s [2024-11-06T08:05:46.959Z] 10803.62 IOPS, 42.20 MiB/s [2024-11-06T08:05:46.959Z] Received shutdown signal, test time was about 26.912872 seconds 00:29:56.846 00:29:56.846 Latency(us) 00:29:56.846 [2024-11-06T08:05:46.959Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:56.846 Job: Nvme0n1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:29:56.846 Verification LBA range: start 0x0 length 0x4000 00:29:56.846 Nvme0n1 : 26.91 10759.22 42.03 0.00 0.00 11877.58 226.99 3019898.88 00:29:56.846 [2024-11-06T08:05:46.959Z] =================================================================================================================== 00:29:56.846 [2024-11-06T08:05:46.959Z] Total : 10759.22 42.03 0.00 0.00 11877.58 226.99 3019898.88 00:29:56.846 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@143 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:29:56.846 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@145 -- # trap - SIGINT SIGTERM EXIT 00:29:56.846 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@147 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:29:56.846 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- host/multipath_status.sh@148 -- # nvmftestfini 00:29:56.846 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@514 -- # nvmfcleanup 00:29:56.846 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@121 -- # sync 00:29:56.846 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:29:56.846 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@124 -- # set +e 00:29:56.846 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@125 -- # for i in {1..20} 00:29:56.846 09:05:46 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:29:57.107 rmmod nvme_tcp 00:29:57.107 rmmod nvme_fabrics 00:29:57.107 rmmod nvme_keyring 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@128 -- # set -e 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@129 -- # return 0 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@515 -- # '[' -n 2018986 ']' 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@516 -- # killprocess 2018986 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@950 -- # '[' -z 2018986 ']' 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # kill -0 2018986 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@955 -- # uname 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2018986 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2018986' 00:29:57.107 killing process with pid 2018986 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@969 -- # kill 2018986 00:29:57.107 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@974 -- # wait 2018986 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@297 -- # iptr 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@789 -- # iptables-restore 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@789 -- # iptables-save 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@302 -- # remove_spdk_ns 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:57.368 09:05:47 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:59.416 09:05:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:29:59.416 00:29:59.416 real 0m40.291s 00:29:59.416 user 1m44.406s 00:29:59.416 sys 0m11.330s 00:29:59.416 09:05:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@1126 -- # xtrace_disable 00:29:59.416 09:05:49 nvmf_tcp.nvmf_host.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:29:59.416 ************************************ 00:29:59.416 END TEST nvmf_host_multipath_status 00:29:59.416 ************************************ 00:29:59.416 09:05:49 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@28 -- # run_test nvmf_discovery_remove_ifc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:29:59.416 09:05:49 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:29:59.416 09:05:49 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:29:59.416 09:05:49 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:29:59.416 ************************************ 00:29:59.416 START TEST nvmf_discovery_remove_ifc 00:29:59.416 ************************************ 00:29:59.417 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:29:59.417 * Looking for test storage... 00:29:59.417 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:29:59.417 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:29:59.417 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1689 -- # lcov --version 00:29:59.417 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@333 -- # local ver1 ver1_l 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@334 -- # local ver2 ver2_l 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@336 -- # IFS=.-: 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@336 -- # read -ra ver1 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@337 -- # IFS=.-: 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@337 -- # read -ra ver2 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@338 -- # local 'op=<' 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@340 -- # ver1_l=2 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@341 -- # ver2_l=1 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@344 -- # case "$op" in 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@345 -- # : 1 00:29:59.725 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@364 -- # (( v = 0 )) 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@365 -- # decimal 1 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@353 -- # local d=1 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@355 -- # echo 1 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@365 -- # ver1[v]=1 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@366 -- # decimal 2 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@353 -- # local d=2 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@355 -- # echo 2 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@366 -- # ver2[v]=2 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@368 -- # return 0 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:29:59.726 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:59.726 --rc genhtml_branch_coverage=1 00:29:59.726 --rc genhtml_function_coverage=1 00:29:59.726 --rc genhtml_legend=1 00:29:59.726 --rc geninfo_all_blocks=1 00:29:59.726 --rc geninfo_unexecuted_blocks=1 00:29:59.726 00:29:59.726 ' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:29:59.726 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:59.726 --rc genhtml_branch_coverage=1 00:29:59.726 --rc genhtml_function_coverage=1 00:29:59.726 --rc genhtml_legend=1 00:29:59.726 --rc geninfo_all_blocks=1 00:29:59.726 --rc geninfo_unexecuted_blocks=1 00:29:59.726 00:29:59.726 ' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:29:59.726 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:59.726 --rc genhtml_branch_coverage=1 00:29:59.726 --rc genhtml_function_coverage=1 00:29:59.726 --rc genhtml_legend=1 00:29:59.726 --rc geninfo_all_blocks=1 00:29:59.726 --rc geninfo_unexecuted_blocks=1 00:29:59.726 00:29:59.726 ' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:29:59.726 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:29:59.726 --rc genhtml_branch_coverage=1 00:29:59.726 --rc genhtml_function_coverage=1 00:29:59.726 --rc genhtml_legend=1 00:29:59.726 --rc geninfo_all_blocks=1 00:29:59.726 --rc geninfo_unexecuted_blocks=1 00:29:59.726 00:29:59.726 ' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@7 -- # uname -s 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@15 -- # shopt -s extglob 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@5 -- # export PATH 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@51 -- # : 0 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:29:59.726 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@55 -- # have_pci_nics=0 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@14 -- # '[' tcp == rdma ']' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@19 -- # discovery_port=8009 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@20 -- # discovery_nqn=nqn.2014-08.org.nvmexpress.discovery 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@23 -- # nqn=nqn.2016-06.io.spdk:cnode 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@25 -- # host_nqn=nqn.2021-12.io.spdk:test 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@26 -- # host_sock=/tmp/host.sock 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@39 -- # nvmftestinit 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@474 -- # prepare_net_devs 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@436 -- # local -g is_hw=no 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@438 -- # remove_spdk_ns 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@309 -- # xtrace_disable 00:29:59.726 09:05:49 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@315 -- # pci_devs=() 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@319 -- # net_devs=() 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@320 -- # e810=() 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@320 -- # local -ga e810 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@321 -- # x722=() 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@321 -- # local -ga x722 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@322 -- # mlx=() 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@322 -- # local -ga mlx 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:07.876 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:30:07.877 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:30:07.877 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:30:07.877 Found net devices under 0000:4b:00.0: cvl_0_0 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:30:07.877 Found net devices under 0000:4b:00.1: cvl_0_1 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@440 -- # is_hw=yes 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:30:07.877 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:30:07.877 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.679 ms 00:30:07.877 00:30:07.877 --- 10.0.0.2 ping statistics --- 00:30:07.877 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:07.877 rtt min/avg/max/mdev = 0.679/0.679/0.679/0.000 ms 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:30:07.877 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:30:07.877 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.276 ms 00:30:07.877 00:30:07.877 --- 10.0.0.1 ping statistics --- 00:30:07.877 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:07.877 rtt min/avg/max/mdev = 0.276/0.276/0.276/0.000 ms 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@448 -- # return 0 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@40 -- # nvmfappstart -m 0x2 00:30:07.877 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@724 -- # xtrace_disable 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@507 -- # nvmfpid=2029243 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@508 -- # waitforlisten 2029243 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@831 -- # '[' -z 2029243 ']' 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:07.878 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:07.878 09:05:56 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:07.878 [2024-11-06 09:05:57.024013] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:30:07.878 [2024-11-06 09:05:57.024078] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:30:07.878 [2024-11-06 09:05:57.124662] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:07.878 [2024-11-06 09:05:57.175059] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:30:07.878 [2024-11-06 09:05:57.175113] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:30:07.878 [2024-11-06 09:05:57.175121] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:30:07.878 [2024-11-06 09:05:57.175129] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:30:07.878 [2024-11-06 09:05:57.175135] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:30:07.878 [2024-11-06 09:05:57.175911] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@864 -- # return 0 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@730 -- # xtrace_disable 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@43 -- # rpc_cmd 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:07.878 [2024-11-06 09:05:57.898271] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:30:07.878 [2024-11-06 09:05:57.906580] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:30:07.878 null0 00:30:07.878 [2024-11-06 09:05:57.938485] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@59 -- # hostpid=2029563 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock --wait-for-rpc -L bdev_nvme 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@60 -- # waitforlisten 2029563 /tmp/host.sock 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@831 -- # '[' -z 2029563 ']' 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@835 -- # local rpc_addr=/tmp/host.sock 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:30:07.878 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:07.878 09:05:57 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:08.143 [2024-11-06 09:05:58.029352] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:30:08.143 [2024-11-06 09:05:58.029426] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2029563 ] 00:30:08.143 [2024-11-06 09:05:58.104893] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:30:08.143 [2024-11-06 09:05:58.146680] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@864 -- # return 0 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@62 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@65 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_set_options -e 1 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@66 -- # rpc_cmd -s /tmp/host.sock framework_start_init 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@69 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test --ctrlr-loss-timeout-sec 2 --reconnect-delay-sec 1 --fast-io-fail-timeout-sec 1 --wait-for-attach 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:09.084 09:05:58 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:10.025 [2024-11-06 09:05:59.957693] bdev_nvme.c:7291:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:30:10.025 [2024-11-06 09:05:59.957713] bdev_nvme.c:7377:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:30:10.025 [2024-11-06 09:05:59.957726] bdev_nvme.c:7254:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:30:10.025 [2024-11-06 09:06:00.087195] bdev_nvme.c:7220:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:30:10.286 [2024-11-06 09:06:00.146079] bdev_nvme.c:5582:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr was created to 10.0.0.2:4420 00:30:10.286 [2024-11-06 09:06:00.147269] bdev_nvme.c:1963:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Connecting qpair 0x21e93b0:1 started. 00:30:10.286 [2024-11-06 09:06:00.148833] bdev_nvme.c:8087:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:30:10.286 [2024-11-06 09:06:00.148882] bdev_nvme.c:8087:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:30:10.286 [2024-11-06 09:06:00.148904] bdev_nvme.c:8087:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:30:10.286 [2024-11-06 09:06:00.148918] bdev_nvme.c:7110:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:30:10.286 [2024-11-06 09:06:00.148941] bdev_nvme.c:7069:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@72 -- # wait_for_bdev nvme0n1 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:10.286 [2024-11-06 09:06:00.155967] bdev_nvme.c:1779:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpair 0x21e93b0 was disconnected and freed. delete nvme_qpair. 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != \n\v\m\e\0\n\1 ]] 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@75 -- # ip netns exec cvl_0_0_ns_spdk ip addr del 10.0.0.2/24 dev cvl_0_0 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@76 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 down 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@79 -- # wait_for_bdev '' 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:30:10.286 09:06:00 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:30:11.672 09:06:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:11.672 09:06:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:11.672 09:06:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:11.672 09:06:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:11.672 09:06:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:11.672 09:06:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:11.672 09:06:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:11.672 09:06:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:11.672 09:06:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:30:11.672 09:06:01 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:30:12.613 09:06:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:12.613 09:06:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:12.613 09:06:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:12.613 09:06:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:12.613 09:06:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:12.613 09:06:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:12.613 09:06:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:12.613 09:06:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:12.613 09:06:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:30:12.613 09:06:02 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:30:13.557 09:06:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:13.557 09:06:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:13.557 09:06:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:13.557 09:06:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:13.557 09:06:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:13.557 09:06:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:13.557 09:06:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:13.557 09:06:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:13.557 09:06:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:30:13.557 09:06:03 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:30:14.499 09:06:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:14.499 09:06:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:14.499 09:06:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:14.499 09:06:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:14.499 09:06:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:14.499 09:06:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:14.499 09:06:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:14.499 09:06:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:14.499 09:06:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:30:14.499 09:06:04 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:30:15.882 [2024-11-06 09:06:05.589506] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 110: Connection timed out 00:30:15.882 [2024-11-06 09:06:05.589548] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:30:15.882 [2024-11-06 09:06:05.589561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:15.882 [2024-11-06 09:06:05.589571] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:30:15.882 [2024-11-06 09:06:05.589579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:15.882 [2024-11-06 09:06:05.589587] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:30:15.882 [2024-11-06 09:06:05.589594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:15.882 [2024-11-06 09:06:05.589602] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:30:15.883 [2024-11-06 09:06:05.589609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:15.883 [2024-11-06 09:06:05.589618] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:30:15.883 [2024-11-06 09:06:05.589625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:15.883 [2024-11-06 09:06:05.589637] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x21c5c00 is same with the state(6) to be set 00:30:15.883 09:06:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:15.883 09:06:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:15.883 [2024-11-06 09:06:05.599525] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x21c5c00 (9): Bad file descriptor 00:30:15.883 09:06:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:15.883 09:06:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:15.883 09:06:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:15.883 09:06:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:15.883 09:06:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:15.883 [2024-11-06 09:06:05.609575] bdev_nvme.c:2536:bdev_nvme_reset_destroy_qpairs: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Delete qpairs for reset. 00:30:15.883 [2024-11-06 09:06:05.609590] bdev_nvme.c:2524:bdev_nvme_reset_destroy_qpair_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] qpairs were deleted. 00:30:15.883 [2024-11-06 09:06:05.609595] bdev_nvme.c:2112:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:30:15.883 [2024-11-06 09:06:05.609601] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:30:15.883 [2024-11-06 09:06:05.609621] bdev_nvme.c:2508:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:30:15.883 09:06:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:15.883 09:06:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:30:15.883 09:06:05 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:30:16.824 [2024-11-06 09:06:06.644787] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 110 00:30:16.824 [2024-11-06 09:06:06.644833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x21c5c00 with addr=10.0.0.2, port=4420 00:30:16.824 [2024-11-06 09:06:06.644848] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x21c5c00 is same with the state(6) to be set 00:30:16.824 [2024-11-06 09:06:06.644879] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x21c5c00 (9): Bad file descriptor 00:30:16.824 [2024-11-06 09:06:06.644932] bdev_nvme.c:3152:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] Unable to perform failover, already in progress. 00:30:16.824 [2024-11-06 09:06:06.644957] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:30:16.824 [2024-11-06 09:06:06.644965] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:30:16.824 [2024-11-06 09:06:06.644974] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:30:16.824 [2024-11-06 09:06:06.644983] bdev_nvme.c:2498:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:30:16.824 [2024-11-06 09:06:06.644989] bdev_nvme.c:2305:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:30:16.824 [2024-11-06 09:06:06.645005] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:30:16.824 [2024-11-06 09:06:06.645014] bdev_nvme.c:2112:nvme_ctrlr_disconnect: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start disconnecting ctrlr. 00:30:16.824 [2024-11-06 09:06:06.645020] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:30:16.824 09:06:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:16.824 09:06:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:16.824 09:06:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:16.824 09:06:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:16.824 09:06:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:16.824 09:06:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:16.824 09:06:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:16.824 09:06:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:16.824 09:06:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:30:16.824 09:06:06 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:30:17.766 [2024-11-06 09:06:07.647394] bdev_nvme.c:2508:bdev_nvme_reconnect_ctrlr: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Start reconnecting ctrlr. 00:30:17.766 [2024-11-06 09:06:07.647417] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] in failed state. 00:30:17.766 [2024-11-06 09:06:07.647430] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Ctrlr is in error state 00:30:17.766 [2024-11-06 09:06:07.647437] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] controller reinitialization failed 00:30:17.766 [2024-11-06 09:06:07.647445] nvme_ctrlr.c:1071:nvme_ctrlr_fail: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] already in failed state 00:30:17.766 [2024-11-06 09:06:07.647452] bdev_nvme.c:2498:bdev_nvme_reconnect_ctrlr_poll: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] ctrlr could not be connected. 00:30:17.766 [2024-11-06 09:06:07.647458] bdev_nvme.c:2305:bdev_nvme_reset_ctrlr_complete: *INFO*: [nqn.2016-06.io.spdk:cnode0, 1] Clear pending resets. 00:30:17.766 [2024-11-06 09:06:07.647470] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 1] Resetting controller failed. 00:30:17.766 [2024-11-06 09:06:07.647490] bdev_nvme.c:7042:remove_discovery_entry: *INFO*: Discovery[10.0.0.2:8009] Remove discovery entry: nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 00:30:17.766 [2024-11-06 09:06:07.647515] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:30:17.766 [2024-11-06 09:06:07.647526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:17.766 [2024-11-06 09:06:07.647538] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:30:17.766 [2024-11-06 09:06:07.647545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:17.766 [2024-11-06 09:06:07.647554] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:30:17.766 [2024-11-06 09:06:07.647562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:17.766 [2024-11-06 09:06:07.647570] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:30:17.766 [2024-11-06 09:06:07.647578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:17.766 [2024-11-06 09:06:07.647586] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:30:17.766 [2024-11-06 09:06:07.647594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:30:17.766 [2024-11-06 09:06:07.647602] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery, 1] in failed state. 00:30:17.766 [2024-11-06 09:06:07.647659] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x21b5340 (9): Bad file descriptor 00:30:17.766 [2024-11-06 09:06:07.648651] nvme_fabric.c: 214:nvme_fabric_prop_get_cmd_async: *ERROR*: Failed to send Property Get fabrics command 00:30:17.766 [2024-11-06 09:06:07.648662] nvme_ctrlr.c:1190:nvme_ctrlr_shutdown_async: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery, 1] Failed to read the CC register 00:30:17.766 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:17.766 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:17.766 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:17.766 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:17.766 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != '' ]] 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@82 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@83 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@86 -- # wait_for_bdev nvme1n1 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:17.767 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.027 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:30:18.027 09:06:07 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:30:18.968 09:06:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:18.968 09:06:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:18.968 09:06:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:18.968 09:06:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:18.968 09:06:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:18.968 09:06:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:18.968 09:06:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:18.968 09:06:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:18.968 09:06:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:30:18.968 09:06:08 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:30:19.910 [2024-11-06 09:06:09.702938] bdev_nvme.c:7291:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:30:19.910 [2024-11-06 09:06:09.702959] bdev_nvme.c:7377:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:30:19.910 [2024-11-06 09:06:09.702972] bdev_nvme.c:7254:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:30:19.910 [2024-11-06 09:06:09.790249] bdev_nvme.c:7220:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme1 00:30:19.910 [2024-11-06 09:06:09.849971] bdev_nvme.c:5582:nvme_ctrlr_create_done: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] ctrlr was created to 10.0.0.2:4420 00:30:19.910 [2024-11-06 09:06:09.850781] bdev_nvme.c:1963:bdev_nvme_create_qpair: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] Connecting qpair 0x21c93a0:1 started. 00:30:19.910 [2024-11-06 09:06:09.851984] bdev_nvme.c:8087:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:30:19.910 [2024-11-06 09:06:09.852017] bdev_nvme.c:8087:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:30:19.910 [2024-11-06 09:06:09.852038] bdev_nvme.c:8087:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:30:19.910 [2024-11-06 09:06:09.852053] bdev_nvme.c:7110:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme1 done 00:30:19.910 [2024-11-06 09:06:09.852061] bdev_nvme.c:7069:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:30:19.910 [2024-11-06 09:06:09.859942] bdev_nvme.c:1779:bdev_nvme_disconnected_qpair_cb: *INFO*: [nqn.2016-06.io.spdk:cnode0, 2] qpair 0x21c93a0 was disconnected and freed. delete nvme_qpair. 00:30:19.910 09:06:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:30:19.910 09:06:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:30:19.910 09:06:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:30:19.910 09:06:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:19.910 09:06:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:30:19.910 09:06:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:19.910 09:06:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:30:19.910 09:06:09 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme1n1 != \n\v\m\e\1\n\1 ]] 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@88 -- # trap - SIGINT SIGTERM EXIT 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@90 -- # killprocess 2029563 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@950 -- # '[' -z 2029563 ']' 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # kill -0 2029563 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@955 -- # uname 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2029563 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2029563' 00:30:20.171 killing process with pid 2029563 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@969 -- # kill 2029563 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@974 -- # wait 2029563 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@91 -- # nvmftestfini 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@514 -- # nvmfcleanup 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@121 -- # sync 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@124 -- # set +e 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@125 -- # for i in {1..20} 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:30:20.171 rmmod nvme_tcp 00:30:20.171 rmmod nvme_fabrics 00:30:20.171 rmmod nvme_keyring 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@128 -- # set -e 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@129 -- # return 0 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@515 -- # '[' -n 2029243 ']' 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@516 -- # killprocess 2029243 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@950 -- # '[' -z 2029243 ']' 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # kill -0 2029243 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@955 -- # uname 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:30:20.171 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2029243 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2029243' 00:30:20.433 killing process with pid 2029243 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@969 -- # kill 2029243 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@974 -- # wait 2029243 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@297 -- # iptr 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@789 -- # iptables-save 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@789 -- # iptables-restore 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@302 -- # remove_spdk_ns 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:20.433 09:06:10 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:30:22.982 00:30:22.982 real 0m23.140s 00:30:22.982 user 0m27.201s 00:30:22.982 sys 0m7.042s 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1126 -- # xtrace_disable 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:30:22.982 ************************************ 00:30:22.982 END TEST nvmf_discovery_remove_ifc 00:30:22.982 ************************************ 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@29 -- # run_test nvmf_identify_kernel_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:30:22.982 ************************************ 00:30:22.982 START TEST nvmf_identify_kernel_target 00:30:22.982 ************************************ 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:30:22.982 * Looking for test storage... 00:30:22.982 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1689 -- # lcov --version 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@336 -- # IFS=.-: 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@336 -- # read -ra ver1 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@337 -- # IFS=.-: 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@337 -- # read -ra ver2 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@338 -- # local 'op=<' 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@340 -- # ver1_l=2 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@341 -- # ver2_l=1 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@344 -- # case "$op" in 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@345 -- # : 1 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@365 -- # decimal 1 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@353 -- # local d=1 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@355 -- # echo 1 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@365 -- # ver1[v]=1 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@366 -- # decimal 2 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@353 -- # local d=2 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@355 -- # echo 2 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@366 -- # ver2[v]=2 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@368 -- # return 0 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:30:22.982 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:22.982 --rc genhtml_branch_coverage=1 00:30:22.982 --rc genhtml_function_coverage=1 00:30:22.982 --rc genhtml_legend=1 00:30:22.982 --rc geninfo_all_blocks=1 00:30:22.982 --rc geninfo_unexecuted_blocks=1 00:30:22.982 00:30:22.982 ' 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:30:22.982 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:22.982 --rc genhtml_branch_coverage=1 00:30:22.982 --rc genhtml_function_coverage=1 00:30:22.982 --rc genhtml_legend=1 00:30:22.982 --rc geninfo_all_blocks=1 00:30:22.982 --rc geninfo_unexecuted_blocks=1 00:30:22.982 00:30:22.982 ' 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:30:22.982 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:22.982 --rc genhtml_branch_coverage=1 00:30:22.982 --rc genhtml_function_coverage=1 00:30:22.982 --rc genhtml_legend=1 00:30:22.982 --rc geninfo_all_blocks=1 00:30:22.982 --rc geninfo_unexecuted_blocks=1 00:30:22.982 00:30:22.982 ' 00:30:22.982 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:30:22.983 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:22.983 --rc genhtml_branch_coverage=1 00:30:22.983 --rc genhtml_function_coverage=1 00:30:22.983 --rc genhtml_legend=1 00:30:22.983 --rc geninfo_all_blocks=1 00:30:22.983 --rc geninfo_unexecuted_blocks=1 00:30:22.983 00:30:22.983 ' 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@7 -- # uname -s 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@15 -- # shopt -s extglob 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@5 -- # export PATH 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@51 -- # : 0 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:30:22.983 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@11 -- # nvmftestinit 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@474 -- # prepare_net_devs 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@436 -- # local -g is_hw=no 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@438 -- # remove_spdk_ns 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@309 -- # xtrace_disable 00:30:22.983 09:06:12 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@10 -- # set +x 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@315 -- # pci_devs=() 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@319 -- # net_devs=() 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@320 -- # e810=() 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@320 -- # local -ga e810 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@321 -- # x722=() 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@321 -- # local -ga x722 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@322 -- # mlx=() 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@322 -- # local -ga mlx 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:30:31.133 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:30:31.133 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:30:31.133 Found net devices under 0000:4b:00.0: cvl_0_0 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:30:31.133 Found net devices under 0000:4b:00.1: cvl_0_1 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@440 -- # is_hw=yes 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:30:31.133 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:30:31.134 09:06:19 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:30:31.134 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:30:31.134 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.673 ms 00:30:31.134 00:30:31.134 --- 10.0.0.2 ping statistics --- 00:30:31.134 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:31.134 rtt min/avg/max/mdev = 0.673/0.673/0.673/0.000 ms 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:30:31.134 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:30:31.134 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.297 ms 00:30:31.134 00:30:31.134 --- 10.0.0.1 ping statistics --- 00:30:31.134 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:31.134 rtt min/avg/max/mdev = 0.297/0.297/0.297/0.000 ms 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@448 -- # return 0 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@13 -- # trap 'nvmftestfini || :; clean_kernel_target' EXIT 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@15 -- # get_main_ns_ip 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@767 -- # local ip 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@15 -- # target_ip=10.0.0.1 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@16 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@658 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@660 -- # nvmet=/sys/kernel/config/nvmet 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@661 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@662 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@663 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@665 -- # local block nvme 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@667 -- # [[ ! -e /sys/module/nvmet ]] 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@668 -- # modprobe nvmet 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@671 -- # [[ -e /sys/kernel/config/nvmet ]] 00:30:31.134 09:06:20 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@673 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:30:33.682 Waiting for block devices as requested 00:30:33.682 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:30:33.682 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:30:33.944 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:30:33.944 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:30:33.944 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:30:34.205 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:30:34.205 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:30:34.205 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:30:34.467 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:30:34.467 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:30:34.467 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:30:34.728 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:30:34.728 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:30:34.728 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:30:34.728 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:30:34.988 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:30:34.988 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@676 -- # for block in /sys/block/nvme* 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@677 -- # [[ -e /sys/block/nvme0n1 ]] 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@678 -- # is_block_zoned nvme0n1 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1646 -- # local device=nvme0n1 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1648 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1649 -- # [[ none != none ]] 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@679 -- # block_in_use nvme0n1 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:30:35.249 No valid GPT data, bailing 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@394 -- # pt= 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- scripts/common.sh@395 -- # return 1 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@679 -- # nvme=/dev/nvme0n1 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@682 -- # [[ -b /dev/nvme0n1 ]] 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@684 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@685 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@691 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@693 -- # echo 1 00:30:35.249 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@694 -- # echo /dev/nvme0n1 00:30:35.250 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@695 -- # echo 1 00:30:35.250 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@697 -- # echo 10.0.0.1 00:30:35.250 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@698 -- # echo tcp 00:30:35.250 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@699 -- # echo 4420 00:30:35.250 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@700 -- # echo ipv4 00:30:35.250 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@703 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:30:35.510 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@706 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -a 10.0.0.1 -t tcp -s 4420 00:30:35.510 00:30:35.510 Discovery Log Number of Records 2, Generation counter 2 00:30:35.510 =====Discovery Log Entry 0====== 00:30:35.510 trtype: tcp 00:30:35.510 adrfam: ipv4 00:30:35.510 subtype: current discovery subsystem 00:30:35.510 treq: not specified, sq flow control disable supported 00:30:35.510 portid: 1 00:30:35.510 trsvcid: 4420 00:30:35.510 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:30:35.510 traddr: 10.0.0.1 00:30:35.510 eflags: none 00:30:35.510 sectype: none 00:30:35.510 =====Discovery Log Entry 1====== 00:30:35.510 trtype: tcp 00:30:35.510 adrfam: ipv4 00:30:35.510 subtype: nvme subsystem 00:30:35.510 treq: not specified, sq flow control disable supported 00:30:35.510 portid: 1 00:30:35.510 trsvcid: 4420 00:30:35.510 subnqn: nqn.2016-06.io.spdk:testnqn 00:30:35.510 traddr: 10.0.0.1 00:30:35.510 eflags: none 00:30:35.510 sectype: none 00:30:35.510 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 00:30:35.510 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' 00:30:35.510 ===================================================== 00:30:35.510 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2014-08.org.nvmexpress.discovery 00:30:35.510 ===================================================== 00:30:35.510 Controller Capabilities/Features 00:30:35.510 ================================ 00:30:35.510 Vendor ID: 0000 00:30:35.510 Subsystem Vendor ID: 0000 00:30:35.510 Serial Number: cf1a0e83c496bd1659f8 00:30:35.510 Model Number: Linux 00:30:35.510 Firmware Version: 6.8.9-20 00:30:35.510 Recommended Arb Burst: 0 00:30:35.510 IEEE OUI Identifier: 00 00 00 00:30:35.510 Multi-path I/O 00:30:35.510 May have multiple subsystem ports: No 00:30:35.510 May have multiple controllers: No 00:30:35.510 Associated with SR-IOV VF: No 00:30:35.510 Max Data Transfer Size: Unlimited 00:30:35.510 Max Number of Namespaces: 0 00:30:35.510 Max Number of I/O Queues: 1024 00:30:35.510 NVMe Specification Version (VS): 1.3 00:30:35.510 NVMe Specification Version (Identify): 1.3 00:30:35.510 Maximum Queue Entries: 1024 00:30:35.510 Contiguous Queues Required: No 00:30:35.510 Arbitration Mechanisms Supported 00:30:35.510 Weighted Round Robin: Not Supported 00:30:35.510 Vendor Specific: Not Supported 00:30:35.510 Reset Timeout: 7500 ms 00:30:35.510 Doorbell Stride: 4 bytes 00:30:35.510 NVM Subsystem Reset: Not Supported 00:30:35.510 Command Sets Supported 00:30:35.510 NVM Command Set: Supported 00:30:35.510 Boot Partition: Not Supported 00:30:35.510 Memory Page Size Minimum: 4096 bytes 00:30:35.510 Memory Page Size Maximum: 4096 bytes 00:30:35.510 Persistent Memory Region: Not Supported 00:30:35.510 Optional Asynchronous Events Supported 00:30:35.510 Namespace Attribute Notices: Not Supported 00:30:35.510 Firmware Activation Notices: Not Supported 00:30:35.510 ANA Change Notices: Not Supported 00:30:35.510 PLE Aggregate Log Change Notices: Not Supported 00:30:35.510 LBA Status Info Alert Notices: Not Supported 00:30:35.510 EGE Aggregate Log Change Notices: Not Supported 00:30:35.511 Normal NVM Subsystem Shutdown event: Not Supported 00:30:35.511 Zone Descriptor Change Notices: Not Supported 00:30:35.511 Discovery Log Change Notices: Supported 00:30:35.511 Controller Attributes 00:30:35.511 128-bit Host Identifier: Not Supported 00:30:35.511 Non-Operational Permissive Mode: Not Supported 00:30:35.511 NVM Sets: Not Supported 00:30:35.511 Read Recovery Levels: Not Supported 00:30:35.511 Endurance Groups: Not Supported 00:30:35.511 Predictable Latency Mode: Not Supported 00:30:35.511 Traffic Based Keep ALive: Not Supported 00:30:35.511 Namespace Granularity: Not Supported 00:30:35.511 SQ Associations: Not Supported 00:30:35.511 UUID List: Not Supported 00:30:35.511 Multi-Domain Subsystem: Not Supported 00:30:35.511 Fixed Capacity Management: Not Supported 00:30:35.511 Variable Capacity Management: Not Supported 00:30:35.511 Delete Endurance Group: Not Supported 00:30:35.511 Delete NVM Set: Not Supported 00:30:35.511 Extended LBA Formats Supported: Not Supported 00:30:35.511 Flexible Data Placement Supported: Not Supported 00:30:35.511 00:30:35.511 Controller Memory Buffer Support 00:30:35.511 ================================ 00:30:35.511 Supported: No 00:30:35.511 00:30:35.511 Persistent Memory Region Support 00:30:35.511 ================================ 00:30:35.511 Supported: No 00:30:35.511 00:30:35.511 Admin Command Set Attributes 00:30:35.511 ============================ 00:30:35.511 Security Send/Receive: Not Supported 00:30:35.511 Format NVM: Not Supported 00:30:35.511 Firmware Activate/Download: Not Supported 00:30:35.511 Namespace Management: Not Supported 00:30:35.511 Device Self-Test: Not Supported 00:30:35.511 Directives: Not Supported 00:30:35.511 NVMe-MI: Not Supported 00:30:35.511 Virtualization Management: Not Supported 00:30:35.511 Doorbell Buffer Config: Not Supported 00:30:35.511 Get LBA Status Capability: Not Supported 00:30:35.511 Command & Feature Lockdown Capability: Not Supported 00:30:35.511 Abort Command Limit: 1 00:30:35.511 Async Event Request Limit: 1 00:30:35.511 Number of Firmware Slots: N/A 00:30:35.511 Firmware Slot 1 Read-Only: N/A 00:30:35.511 Firmware Activation Without Reset: N/A 00:30:35.511 Multiple Update Detection Support: N/A 00:30:35.511 Firmware Update Granularity: No Information Provided 00:30:35.511 Per-Namespace SMART Log: No 00:30:35.511 Asymmetric Namespace Access Log Page: Not Supported 00:30:35.511 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:30:35.511 Command Effects Log Page: Not Supported 00:30:35.511 Get Log Page Extended Data: Supported 00:30:35.511 Telemetry Log Pages: Not Supported 00:30:35.511 Persistent Event Log Pages: Not Supported 00:30:35.511 Supported Log Pages Log Page: May Support 00:30:35.511 Commands Supported & Effects Log Page: Not Supported 00:30:35.511 Feature Identifiers & Effects Log Page:May Support 00:30:35.511 NVMe-MI Commands & Effects Log Page: May Support 00:30:35.511 Data Area 4 for Telemetry Log: Not Supported 00:30:35.511 Error Log Page Entries Supported: 1 00:30:35.511 Keep Alive: Not Supported 00:30:35.511 00:30:35.511 NVM Command Set Attributes 00:30:35.511 ========================== 00:30:35.511 Submission Queue Entry Size 00:30:35.511 Max: 1 00:30:35.511 Min: 1 00:30:35.511 Completion Queue Entry Size 00:30:35.511 Max: 1 00:30:35.511 Min: 1 00:30:35.511 Number of Namespaces: 0 00:30:35.511 Compare Command: Not Supported 00:30:35.511 Write Uncorrectable Command: Not Supported 00:30:35.511 Dataset Management Command: Not Supported 00:30:35.511 Write Zeroes Command: Not Supported 00:30:35.511 Set Features Save Field: Not Supported 00:30:35.511 Reservations: Not Supported 00:30:35.511 Timestamp: Not Supported 00:30:35.511 Copy: Not Supported 00:30:35.511 Volatile Write Cache: Not Present 00:30:35.511 Atomic Write Unit (Normal): 1 00:30:35.511 Atomic Write Unit (PFail): 1 00:30:35.511 Atomic Compare & Write Unit: 1 00:30:35.511 Fused Compare & Write: Not Supported 00:30:35.511 Scatter-Gather List 00:30:35.511 SGL Command Set: Supported 00:30:35.511 SGL Keyed: Not Supported 00:30:35.511 SGL Bit Bucket Descriptor: Not Supported 00:30:35.511 SGL Metadata Pointer: Not Supported 00:30:35.511 Oversized SGL: Not Supported 00:30:35.511 SGL Metadata Address: Not Supported 00:30:35.511 SGL Offset: Supported 00:30:35.511 Transport SGL Data Block: Not Supported 00:30:35.511 Replay Protected Memory Block: Not Supported 00:30:35.511 00:30:35.511 Firmware Slot Information 00:30:35.511 ========================= 00:30:35.511 Active slot: 0 00:30:35.511 00:30:35.511 00:30:35.511 Error Log 00:30:35.511 ========= 00:30:35.511 00:30:35.511 Active Namespaces 00:30:35.511 ================= 00:30:35.511 Discovery Log Page 00:30:35.511 ================== 00:30:35.511 Generation Counter: 2 00:30:35.511 Number of Records: 2 00:30:35.511 Record Format: 0 00:30:35.511 00:30:35.511 Discovery Log Entry 0 00:30:35.511 ---------------------- 00:30:35.511 Transport Type: 3 (TCP) 00:30:35.511 Address Family: 1 (IPv4) 00:30:35.511 Subsystem Type: 3 (Current Discovery Subsystem) 00:30:35.511 Entry Flags: 00:30:35.511 Duplicate Returned Information: 0 00:30:35.511 Explicit Persistent Connection Support for Discovery: 0 00:30:35.511 Transport Requirements: 00:30:35.511 Secure Channel: Not Specified 00:30:35.511 Port ID: 1 (0x0001) 00:30:35.511 Controller ID: 65535 (0xffff) 00:30:35.511 Admin Max SQ Size: 32 00:30:35.511 Transport Service Identifier: 4420 00:30:35.511 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:30:35.511 Transport Address: 10.0.0.1 00:30:35.511 Discovery Log Entry 1 00:30:35.511 ---------------------- 00:30:35.511 Transport Type: 3 (TCP) 00:30:35.511 Address Family: 1 (IPv4) 00:30:35.511 Subsystem Type: 2 (NVM Subsystem) 00:30:35.511 Entry Flags: 00:30:35.511 Duplicate Returned Information: 0 00:30:35.511 Explicit Persistent Connection Support for Discovery: 0 00:30:35.511 Transport Requirements: 00:30:35.511 Secure Channel: Not Specified 00:30:35.511 Port ID: 1 (0x0001) 00:30:35.511 Controller ID: 65535 (0xffff) 00:30:35.511 Admin Max SQ Size: 32 00:30:35.511 Transport Service Identifier: 4420 00:30:35.511 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:testnqn 00:30:35.511 Transport Address: 10.0.0.1 00:30:35.511 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:30:35.772 get_feature(0x01) failed 00:30:35.772 get_feature(0x02) failed 00:30:35.772 get_feature(0x04) failed 00:30:35.772 ===================================================== 00:30:35.772 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:30:35.772 ===================================================== 00:30:35.772 Controller Capabilities/Features 00:30:35.772 ================================ 00:30:35.772 Vendor ID: 0000 00:30:35.772 Subsystem Vendor ID: 0000 00:30:35.772 Serial Number: 381dd6f0f7ce55e23ff4 00:30:35.772 Model Number: SPDK-nqn.2016-06.io.spdk:testnqn 00:30:35.772 Firmware Version: 6.8.9-20 00:30:35.772 Recommended Arb Burst: 6 00:30:35.772 IEEE OUI Identifier: 00 00 00 00:30:35.772 Multi-path I/O 00:30:35.772 May have multiple subsystem ports: Yes 00:30:35.772 May have multiple controllers: Yes 00:30:35.772 Associated with SR-IOV VF: No 00:30:35.772 Max Data Transfer Size: Unlimited 00:30:35.772 Max Number of Namespaces: 1024 00:30:35.772 Max Number of I/O Queues: 128 00:30:35.772 NVMe Specification Version (VS): 1.3 00:30:35.772 NVMe Specification Version (Identify): 1.3 00:30:35.772 Maximum Queue Entries: 1024 00:30:35.772 Contiguous Queues Required: No 00:30:35.772 Arbitration Mechanisms Supported 00:30:35.772 Weighted Round Robin: Not Supported 00:30:35.772 Vendor Specific: Not Supported 00:30:35.772 Reset Timeout: 7500 ms 00:30:35.772 Doorbell Stride: 4 bytes 00:30:35.772 NVM Subsystem Reset: Not Supported 00:30:35.772 Command Sets Supported 00:30:35.772 NVM Command Set: Supported 00:30:35.772 Boot Partition: Not Supported 00:30:35.772 Memory Page Size Minimum: 4096 bytes 00:30:35.772 Memory Page Size Maximum: 4096 bytes 00:30:35.772 Persistent Memory Region: Not Supported 00:30:35.772 Optional Asynchronous Events Supported 00:30:35.772 Namespace Attribute Notices: Supported 00:30:35.772 Firmware Activation Notices: Not Supported 00:30:35.772 ANA Change Notices: Supported 00:30:35.772 PLE Aggregate Log Change Notices: Not Supported 00:30:35.772 LBA Status Info Alert Notices: Not Supported 00:30:35.772 EGE Aggregate Log Change Notices: Not Supported 00:30:35.772 Normal NVM Subsystem Shutdown event: Not Supported 00:30:35.772 Zone Descriptor Change Notices: Not Supported 00:30:35.772 Discovery Log Change Notices: Not Supported 00:30:35.772 Controller Attributes 00:30:35.772 128-bit Host Identifier: Supported 00:30:35.772 Non-Operational Permissive Mode: Not Supported 00:30:35.772 NVM Sets: Not Supported 00:30:35.772 Read Recovery Levels: Not Supported 00:30:35.772 Endurance Groups: Not Supported 00:30:35.772 Predictable Latency Mode: Not Supported 00:30:35.772 Traffic Based Keep ALive: Supported 00:30:35.772 Namespace Granularity: Not Supported 00:30:35.772 SQ Associations: Not Supported 00:30:35.772 UUID List: Not Supported 00:30:35.772 Multi-Domain Subsystem: Not Supported 00:30:35.772 Fixed Capacity Management: Not Supported 00:30:35.772 Variable Capacity Management: Not Supported 00:30:35.772 Delete Endurance Group: Not Supported 00:30:35.772 Delete NVM Set: Not Supported 00:30:35.772 Extended LBA Formats Supported: Not Supported 00:30:35.772 Flexible Data Placement Supported: Not Supported 00:30:35.772 00:30:35.772 Controller Memory Buffer Support 00:30:35.772 ================================ 00:30:35.772 Supported: No 00:30:35.772 00:30:35.772 Persistent Memory Region Support 00:30:35.772 ================================ 00:30:35.772 Supported: No 00:30:35.772 00:30:35.772 Admin Command Set Attributes 00:30:35.772 ============================ 00:30:35.772 Security Send/Receive: Not Supported 00:30:35.772 Format NVM: Not Supported 00:30:35.772 Firmware Activate/Download: Not Supported 00:30:35.772 Namespace Management: Not Supported 00:30:35.772 Device Self-Test: Not Supported 00:30:35.772 Directives: Not Supported 00:30:35.772 NVMe-MI: Not Supported 00:30:35.772 Virtualization Management: Not Supported 00:30:35.772 Doorbell Buffer Config: Not Supported 00:30:35.772 Get LBA Status Capability: Not Supported 00:30:35.772 Command & Feature Lockdown Capability: Not Supported 00:30:35.772 Abort Command Limit: 4 00:30:35.772 Async Event Request Limit: 4 00:30:35.772 Number of Firmware Slots: N/A 00:30:35.772 Firmware Slot 1 Read-Only: N/A 00:30:35.772 Firmware Activation Without Reset: N/A 00:30:35.773 Multiple Update Detection Support: N/A 00:30:35.773 Firmware Update Granularity: No Information Provided 00:30:35.773 Per-Namespace SMART Log: Yes 00:30:35.773 Asymmetric Namespace Access Log Page: Supported 00:30:35.773 ANA Transition Time : 10 sec 00:30:35.773 00:30:35.773 Asymmetric Namespace Access Capabilities 00:30:35.773 ANA Optimized State : Supported 00:30:35.773 ANA Non-Optimized State : Supported 00:30:35.773 ANA Inaccessible State : Supported 00:30:35.773 ANA Persistent Loss State : Supported 00:30:35.773 ANA Change State : Supported 00:30:35.773 ANAGRPID is not changed : No 00:30:35.773 Non-Zero ANAGRPID for NS Mgmt Cmd : Not Supported 00:30:35.773 00:30:35.773 ANA Group Identifier Maximum : 128 00:30:35.773 Number of ANA Group Identifiers : 128 00:30:35.773 Max Number of Allowed Namespaces : 1024 00:30:35.773 Subsystem NQN: nqn.2016-06.io.spdk:testnqn 00:30:35.773 Command Effects Log Page: Supported 00:30:35.773 Get Log Page Extended Data: Supported 00:30:35.773 Telemetry Log Pages: Not Supported 00:30:35.773 Persistent Event Log Pages: Not Supported 00:30:35.773 Supported Log Pages Log Page: May Support 00:30:35.773 Commands Supported & Effects Log Page: Not Supported 00:30:35.773 Feature Identifiers & Effects Log Page:May Support 00:30:35.773 NVMe-MI Commands & Effects Log Page: May Support 00:30:35.773 Data Area 4 for Telemetry Log: Not Supported 00:30:35.773 Error Log Page Entries Supported: 128 00:30:35.773 Keep Alive: Supported 00:30:35.773 Keep Alive Granularity: 1000 ms 00:30:35.773 00:30:35.773 NVM Command Set Attributes 00:30:35.773 ========================== 00:30:35.773 Submission Queue Entry Size 00:30:35.773 Max: 64 00:30:35.773 Min: 64 00:30:35.773 Completion Queue Entry Size 00:30:35.773 Max: 16 00:30:35.773 Min: 16 00:30:35.773 Number of Namespaces: 1024 00:30:35.773 Compare Command: Not Supported 00:30:35.773 Write Uncorrectable Command: Not Supported 00:30:35.773 Dataset Management Command: Supported 00:30:35.773 Write Zeroes Command: Supported 00:30:35.773 Set Features Save Field: Not Supported 00:30:35.773 Reservations: Not Supported 00:30:35.773 Timestamp: Not Supported 00:30:35.773 Copy: Not Supported 00:30:35.773 Volatile Write Cache: Present 00:30:35.773 Atomic Write Unit (Normal): 1 00:30:35.773 Atomic Write Unit (PFail): 1 00:30:35.773 Atomic Compare & Write Unit: 1 00:30:35.773 Fused Compare & Write: Not Supported 00:30:35.773 Scatter-Gather List 00:30:35.773 SGL Command Set: Supported 00:30:35.773 SGL Keyed: Not Supported 00:30:35.773 SGL Bit Bucket Descriptor: Not Supported 00:30:35.773 SGL Metadata Pointer: Not Supported 00:30:35.773 Oversized SGL: Not Supported 00:30:35.773 SGL Metadata Address: Not Supported 00:30:35.773 SGL Offset: Supported 00:30:35.773 Transport SGL Data Block: Not Supported 00:30:35.773 Replay Protected Memory Block: Not Supported 00:30:35.773 00:30:35.773 Firmware Slot Information 00:30:35.773 ========================= 00:30:35.773 Active slot: 0 00:30:35.773 00:30:35.773 Asymmetric Namespace Access 00:30:35.773 =========================== 00:30:35.773 Change Count : 0 00:30:35.773 Number of ANA Group Descriptors : 1 00:30:35.773 ANA Group Descriptor : 0 00:30:35.773 ANA Group ID : 1 00:30:35.773 Number of NSID Values : 1 00:30:35.773 Change Count : 0 00:30:35.773 ANA State : 1 00:30:35.773 Namespace Identifier : 1 00:30:35.773 00:30:35.773 Commands Supported and Effects 00:30:35.773 ============================== 00:30:35.773 Admin Commands 00:30:35.773 -------------- 00:30:35.773 Get Log Page (02h): Supported 00:30:35.773 Identify (06h): Supported 00:30:35.773 Abort (08h): Supported 00:30:35.773 Set Features (09h): Supported 00:30:35.773 Get Features (0Ah): Supported 00:30:35.773 Asynchronous Event Request (0Ch): Supported 00:30:35.773 Keep Alive (18h): Supported 00:30:35.773 I/O Commands 00:30:35.773 ------------ 00:30:35.773 Flush (00h): Supported 00:30:35.773 Write (01h): Supported LBA-Change 00:30:35.773 Read (02h): Supported 00:30:35.773 Write Zeroes (08h): Supported LBA-Change 00:30:35.773 Dataset Management (09h): Supported 00:30:35.773 00:30:35.773 Error Log 00:30:35.773 ========= 00:30:35.773 Entry: 0 00:30:35.773 Error Count: 0x3 00:30:35.773 Submission Queue Id: 0x0 00:30:35.773 Command Id: 0x5 00:30:35.773 Phase Bit: 0 00:30:35.773 Status Code: 0x2 00:30:35.773 Status Code Type: 0x0 00:30:35.773 Do Not Retry: 1 00:30:35.773 Error Location: 0x28 00:30:35.773 LBA: 0x0 00:30:35.773 Namespace: 0x0 00:30:35.773 Vendor Log Page: 0x0 00:30:35.773 ----------- 00:30:35.773 Entry: 1 00:30:35.773 Error Count: 0x2 00:30:35.773 Submission Queue Id: 0x0 00:30:35.773 Command Id: 0x5 00:30:35.773 Phase Bit: 0 00:30:35.773 Status Code: 0x2 00:30:35.773 Status Code Type: 0x0 00:30:35.773 Do Not Retry: 1 00:30:35.773 Error Location: 0x28 00:30:35.773 LBA: 0x0 00:30:35.773 Namespace: 0x0 00:30:35.773 Vendor Log Page: 0x0 00:30:35.773 ----------- 00:30:35.773 Entry: 2 00:30:35.773 Error Count: 0x1 00:30:35.773 Submission Queue Id: 0x0 00:30:35.773 Command Id: 0x4 00:30:35.773 Phase Bit: 0 00:30:35.773 Status Code: 0x2 00:30:35.773 Status Code Type: 0x0 00:30:35.773 Do Not Retry: 1 00:30:35.773 Error Location: 0x28 00:30:35.773 LBA: 0x0 00:30:35.773 Namespace: 0x0 00:30:35.773 Vendor Log Page: 0x0 00:30:35.773 00:30:35.773 Number of Queues 00:30:35.773 ================ 00:30:35.773 Number of I/O Submission Queues: 128 00:30:35.773 Number of I/O Completion Queues: 128 00:30:35.773 00:30:35.773 ZNS Specific Controller Data 00:30:35.773 ============================ 00:30:35.773 Zone Append Size Limit: 0 00:30:35.773 00:30:35.773 00:30:35.773 Active Namespaces 00:30:35.773 ================= 00:30:35.773 get_feature(0x05) failed 00:30:35.773 Namespace ID:1 00:30:35.773 Command Set Identifier: NVM (00h) 00:30:35.773 Deallocate: Supported 00:30:35.773 Deallocated/Unwritten Error: Not Supported 00:30:35.773 Deallocated Read Value: Unknown 00:30:35.773 Deallocate in Write Zeroes: Not Supported 00:30:35.773 Deallocated Guard Field: 0xFFFF 00:30:35.773 Flush: Supported 00:30:35.773 Reservation: Not Supported 00:30:35.773 Namespace Sharing Capabilities: Multiple Controllers 00:30:35.773 Size (in LBAs): 3750748848 (1788GiB) 00:30:35.773 Capacity (in LBAs): 3750748848 (1788GiB) 00:30:35.773 Utilization (in LBAs): 3750748848 (1788GiB) 00:30:35.773 UUID: 351998ad-1915-4de7-90ec-e2fd5c90b0ee 00:30:35.773 Thin Provisioning: Not Supported 00:30:35.773 Per-NS Atomic Units: Yes 00:30:35.773 Atomic Write Unit (Normal): 8 00:30:35.773 Atomic Write Unit (PFail): 8 00:30:35.773 Preferred Write Granularity: 8 00:30:35.773 Atomic Compare & Write Unit: 8 00:30:35.773 Atomic Boundary Size (Normal): 0 00:30:35.773 Atomic Boundary Size (PFail): 0 00:30:35.773 Atomic Boundary Offset: 0 00:30:35.773 NGUID/EUI64 Never Reused: No 00:30:35.773 ANA group ID: 1 00:30:35.773 Namespace Write Protected: No 00:30:35.773 Number of LBA Formats: 1 00:30:35.773 Current LBA Format: LBA Format #00 00:30:35.773 LBA Format #00: Data Size: 512 Metadata Size: 0 00:30:35.773 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@1 -- # nvmftestfini 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@514 -- # nvmfcleanup 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@121 -- # sync 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@124 -- # set +e 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:30:35.773 rmmod nvme_tcp 00:30:35.773 rmmod nvme_fabrics 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@128 -- # set -e 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@129 -- # return 0 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@297 -- # iptr 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@789 -- # iptables-save 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@789 -- # iptables-restore 00:30:35.773 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:30:35.774 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:30:35.774 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:35.774 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:35.774 09:06:25 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:37.684 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:30:37.684 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@1 -- # clean_kernel_target 00:30:37.684 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@710 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:30:37.684 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@712 -- # echo 0 00:30:37.684 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@714 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:30:37.684 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@715 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:30:37.684 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@716 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:30:37.945 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:30:37.945 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@719 -- # modules=(/sys/module/nvmet/holders/*) 00:30:37.945 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@721 -- # modprobe -r nvmet_tcp nvmet 00:30:37.945 09:06:27 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- nvmf/common.sh@724 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:30:41.264 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:30:41.264 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:30:41.264 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:30:41.264 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:30:41.264 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:30:41.264 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:30:41.264 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:30:41.264 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:30:41.264 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:30:41.264 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:30:41.524 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:30:41.524 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:30:41.524 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:30:41.524 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:30:41.524 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:30:41.524 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:30:41.524 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:30:41.786 00:30:41.786 real 0m19.237s 00:30:41.786 user 0m5.341s 00:30:41.786 sys 0m11.034s 00:30:41.786 09:06:31 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@1126 -- # xtrace_disable 00:30:41.786 09:06:31 nvmf_tcp.nvmf_host.nvmf_identify_kernel_target -- common/autotest_common.sh@10 -- # set +x 00:30:41.786 ************************************ 00:30:41.786 END TEST nvmf_identify_kernel_target 00:30:41.786 ************************************ 00:30:41.786 09:06:31 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@30 -- # run_test nvmf_auth_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:30:41.786 09:06:31 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:30:41.786 09:06:31 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:30:41.786 09:06:31 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:30:42.047 ************************************ 00:30:42.047 START TEST nvmf_auth_host 00:30:42.047 ************************************ 00:30:42.047 09:06:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:30:42.047 * Looking for test storage... 00:30:42.047 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1689 -- # lcov --version 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@333 -- # local ver1 ver1_l 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@334 -- # local ver2 ver2_l 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@336 -- # IFS=.-: 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@336 -- # read -ra ver1 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@337 -- # IFS=.-: 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@337 -- # read -ra ver2 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@338 -- # local 'op=<' 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@340 -- # ver1_l=2 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@341 -- # ver2_l=1 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@344 -- # case "$op" in 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@345 -- # : 1 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@364 -- # (( v = 0 )) 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@365 -- # decimal 1 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@353 -- # local d=1 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@355 -- # echo 1 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@365 -- # ver1[v]=1 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@366 -- # decimal 2 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@353 -- # local d=2 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@355 -- # echo 2 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@366 -- # ver2[v]=2 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@368 -- # return 0 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:30:42.047 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:42.047 --rc genhtml_branch_coverage=1 00:30:42.047 --rc genhtml_function_coverage=1 00:30:42.047 --rc genhtml_legend=1 00:30:42.047 --rc geninfo_all_blocks=1 00:30:42.047 --rc geninfo_unexecuted_blocks=1 00:30:42.047 00:30:42.047 ' 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:30:42.047 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:42.047 --rc genhtml_branch_coverage=1 00:30:42.047 --rc genhtml_function_coverage=1 00:30:42.047 --rc genhtml_legend=1 00:30:42.047 --rc geninfo_all_blocks=1 00:30:42.047 --rc geninfo_unexecuted_blocks=1 00:30:42.047 00:30:42.047 ' 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:30:42.047 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:42.047 --rc genhtml_branch_coverage=1 00:30:42.047 --rc genhtml_function_coverage=1 00:30:42.047 --rc genhtml_legend=1 00:30:42.047 --rc geninfo_all_blocks=1 00:30:42.047 --rc geninfo_unexecuted_blocks=1 00:30:42.047 00:30:42.047 ' 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:30:42.047 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:30:42.047 --rc genhtml_branch_coverage=1 00:30:42.047 --rc genhtml_function_coverage=1 00:30:42.047 --rc genhtml_legend=1 00:30:42.047 --rc geninfo_all_blocks=1 00:30:42.047 --rc geninfo_unexecuted_blocks=1 00:30:42.047 00:30:42.047 ' 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@7 -- # uname -s 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:30:42.047 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@15 -- # shopt -s extglob 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@5 -- # export PATH 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@51 -- # : 0 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:30:42.048 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:30:42.308 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@55 -- # have_pci_nics=0 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@16 -- # dhgroups=("ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@17 -- # subnqn=nqn.2024-02.io.spdk:cnode0 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@18 -- # hostnqn=nqn.2024-02.io.spdk:host0 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@19 -- # nvmet_subsys=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@20 -- # nvmet_host=/sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@21 -- # keys=() 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@21 -- # ckeys=() 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@68 -- # nvmftestinit 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@474 -- # prepare_net_devs 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@436 -- # local -g is_hw=no 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@438 -- # remove_spdk_ns 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:30:42.308 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:30:42.309 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:30:42.309 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:30:42.309 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@309 -- # xtrace_disable 00:30:42.309 09:06:32 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@315 -- # pci_devs=() 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@315 -- # local -a pci_devs 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@316 -- # pci_net_devs=() 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@317 -- # pci_drivers=() 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@317 -- # local -A pci_drivers 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@319 -- # net_devs=() 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@319 -- # local -ga net_devs 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@320 -- # e810=() 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@320 -- # local -ga e810 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@321 -- # x722=() 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@321 -- # local -ga x722 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@322 -- # mlx=() 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@322 -- # local -ga mlx 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:30:50.448 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:30:50.448 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:30:50.448 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:30:50.449 Found net devices under 0000:4b:00.0: cvl_0_0 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@416 -- # [[ up == up ]] 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:30:50.449 Found net devices under 0000:4b:00.1: cvl_0_1 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@440 -- # is_hw=yes 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:30:50.449 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:30:50.449 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.612 ms 00:30:50.449 00:30:50.449 --- 10.0.0.2 ping statistics --- 00:30:50.449 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:50.449 rtt min/avg/max/mdev = 0.612/0.612/0.612/0.000 ms 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:30:50.449 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:30:50.449 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.269 ms 00:30:50.449 00:30:50.449 --- 10.0.0.1 ping statistics --- 00:30:50.449 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:30:50.449 rtt min/avg/max/mdev = 0.269/0.269/0.269/0.000 ms 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@448 -- # return 0 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@69 -- # nvmfappstart -L nvme_auth 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@724 -- # xtrace_disable 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@507 -- # nvmfpid=2044323 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@508 -- # waitforlisten 2044323 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvme_auth 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@831 -- # '[' -z 2044323 ']' 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:50.449 09:06:39 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@864 -- # return 0 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@730 -- # xtrace_disable 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@70 -- # trap 'cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log; cleanup' SIGINT SIGTERM EXIT 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # gen_dhchap_key null 32 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=null 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=32 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=1c875d50cd1d9f234fd63dcf78f88310 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-null.XXX 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-null.s1J 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key 1c875d50cd1d9f234fd63dcf78f88310 0 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 1c875d50cd1d9f234fd63dcf78f88310 0 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=1c875d50cd1d9f234fd63dcf78f88310 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=0 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-null.s1J 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-null.s1J 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # keys[0]=/tmp/spdk.key-null.s1J 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # gen_dhchap_key sha512 64 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha512 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=64 00:30:50.449 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 32 /dev/urandom 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=ca1a630329c063c4495fd38433a4aa0cb5918e38a65261678e95ea03394298f1 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha512.XXX 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha512.FoR 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key ca1a630329c063c4495fd38433a4aa0cb5918e38a65261678e95ea03394298f1 3 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 ca1a630329c063c4495fd38433a4aa0cb5918e38a65261678e95ea03394298f1 3 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=ca1a630329c063c4495fd38433a4aa0cb5918e38a65261678e95ea03394298f1 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=3 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha512.FoR 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha512.FoR 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@73 -- # ckeys[0]=/tmp/spdk.key-sha512.FoR 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # gen_dhchap_key null 48 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=null 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=48 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=981a3cc398aaf13c6d3008cb39752f8ea73a782dbf2295a8 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-null.XXX 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-null.Igb 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key 981a3cc398aaf13c6d3008cb39752f8ea73a782dbf2295a8 0 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 981a3cc398aaf13c6d3008cb39752f8ea73a782dbf2295a8 0 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=981a3cc398aaf13c6d3008cb39752f8ea73a782dbf2295a8 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=0 00:30:50.450 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-null.Igb 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-null.Igb 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # keys[1]=/tmp/spdk.key-null.Igb 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # gen_dhchap_key sha384 48 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha384 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=48 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=8b0049c32135d712fbfed2d21ed1d0b091795fed0fa93460 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha384.XXX 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha384.fZ8 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key 8b0049c32135d712fbfed2d21ed1d0b091795fed0fa93460 2 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 8b0049c32135d712fbfed2d21ed1d0b091795fed0fa93460 2 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=8b0049c32135d712fbfed2d21ed1d0b091795fed0fa93460 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=2 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha384.fZ8 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha384.fZ8 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@74 -- # ckeys[1]=/tmp/spdk.key-sha384.fZ8 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # gen_dhchap_key sha256 32 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha256 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=32 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=fd55bff7d520dbbb95f94118b6039052 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha256.XXX 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha256.knh 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key fd55bff7d520dbbb95f94118b6039052 1 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 fd55bff7d520dbbb95f94118b6039052 1 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=fd55bff7d520dbbb95f94118b6039052 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=1 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha256.knh 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha256.knh 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # keys[2]=/tmp/spdk.key-sha256.knh 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # gen_dhchap_key sha256 32 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha256 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=32 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=66a94512b25db703b258a5587984a241 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha256.XXX 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha256.lFp 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key 66a94512b25db703b258a5587984a241 1 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 66a94512b25db703b258a5587984a241 1 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=66a94512b25db703b258a5587984a241 00:30:50.711 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=1 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha256.lFp 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha256.lFp 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@75 -- # ckeys[2]=/tmp/spdk.key-sha256.lFp 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # gen_dhchap_key sha384 48 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha384 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=48 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 24 /dev/urandom 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=ab9b12cf1cc6aa02ab34fea25499d1318e6160709dde1c28 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha384.XXX 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha384.bJm 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key ab9b12cf1cc6aa02ab34fea25499d1318e6160709dde1c28 2 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 ab9b12cf1cc6aa02ab34fea25499d1318e6160709dde1c28 2 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=ab9b12cf1cc6aa02ab34fea25499d1318e6160709dde1c28 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=2 00:30:50.712 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha384.bJm 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha384.bJm 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # keys[3]=/tmp/spdk.key-sha384.bJm 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # gen_dhchap_key null 32 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=null 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=32 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 16 /dev/urandom 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=1e1573b7f296fb77f655ada58c6f2321 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-null.XXX 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-null.B4P 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key 1e1573b7f296fb77f655ada58c6f2321 0 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 1e1573b7f296fb77f655ada58c6f2321 0 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=1e1573b7f296fb77f655ada58c6f2321 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=0 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-null.B4P 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-null.B4P 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@76 -- # ckeys[3]=/tmp/spdk.key-null.B4P 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # gen_dhchap_key sha512 64 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@749 -- # local digest len file key 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@750 -- # local -A digests 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # digest=sha512 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@752 -- # len=64 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # xxd -p -c0 -l 32 /dev/urandom 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@753 -- # key=ff29873b2aba7c455bc40485c47fd2fe8cf84055235b2a0358dc08a6913b4487 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # mktemp -t spdk.key-sha512.XXX 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@754 -- # file=/tmp/spdk.key-sha512.AYV 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@755 -- # format_dhchap_key ff29873b2aba7c455bc40485c47fd2fe8cf84055235b2a0358dc08a6913b4487 3 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@745 -- # format_key DHHC-1 ff29873b2aba7c455bc40485c47fd2fe8cf84055235b2a0358dc08a6913b4487 3 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@728 -- # local prefix key digest 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # prefix=DHHC-1 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # key=ff29873b2aba7c455bc40485c47fd2fe8cf84055235b2a0358dc08a6913b4487 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@730 -- # digest=3 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@731 -- # python - 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@756 -- # chmod 0600 /tmp/spdk.key-sha512.AYV 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@758 -- # echo /tmp/spdk.key-sha512.AYV 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # keys[4]=/tmp/spdk.key-sha512.AYV 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@77 -- # ckeys[4]= 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@79 -- # waitforlisten 2044323 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@831 -- # '[' -z 2044323 ']' 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@836 -- # local max_retries=100 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:30:50.972 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@840 -- # xtrace_disable 00:30:50.972 09:06:40 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@864 -- # return 0 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.s1J 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha512.FoR ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.FoR 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-null.Igb 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha384.fZ8 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.fZ8 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha256.knh 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha256.lFp ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.lFp 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha384.bJm 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-null.B4P ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey3 /tmp/spdk.key-null.B4P 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key4 /tmp/spdk.key-sha512.AYV 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n '' ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@85 -- # nvmet_auth_init 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@35 -- # get_main_ns_ip 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@35 -- # configure_kernel_target nqn.2024-02.io.spdk:cnode0 10.0.0.1 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@658 -- # local kernel_name=nqn.2024-02.io.spdk:cnode0 kernel_target_ip=10.0.0.1 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@660 -- # nvmet=/sys/kernel/config/nvmet 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@661 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@662 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@663 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@665 -- # local block nvme 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@667 -- # [[ ! -e /sys/module/nvmet ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@668 -- # modprobe nvmet 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@671 -- # [[ -e /sys/kernel/config/nvmet ]] 00:30:51.233 09:06:41 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@673 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:30:54.529 Waiting for block devices as requested 00:30:54.529 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:30:54.788 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:30:54.788 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:30:54.788 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:30:54.788 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:30:55.048 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:30:55.048 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:30:55.048 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:30:55.307 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:30:55.307 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:30:55.568 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:30:55.568 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:30:55.568 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:30:55.568 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:30:55.829 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:30:55.829 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:30:55.829 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@676 -- # for block in /sys/block/nvme* 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@677 -- # [[ -e /sys/block/nvme0n1 ]] 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@678 -- # is_block_zoned nvme0n1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1646 -- # local device=nvme0n1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1648 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1649 -- # [[ none != none ]] 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@679 -- # block_in_use nvme0n1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:30:56.770 No valid GPT data, bailing 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@394 -- # pt= 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- scripts/common.sh@395 -- # return 1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@679 -- # nvme=/dev/nvme0n1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@682 -- # [[ -b /dev/nvme0n1 ]] 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@684 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@685 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@691 -- # echo SPDK-nqn.2024-02.io.spdk:cnode0 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@693 -- # echo 1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@694 -- # echo /dev/nvme0n1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@695 -- # echo 1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@697 -- # echo 10.0.0.1 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@698 -- # echo tcp 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@699 -- # echo 4420 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@700 -- # echo ipv4 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@703 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 /sys/kernel/config/nvmet/ports/1/subsystems/ 00:30:56.770 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@706 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -a 10.0.0.1 -t tcp -s 4420 00:30:57.031 00:30:57.031 Discovery Log Number of Records 2, Generation counter 2 00:30:57.031 =====Discovery Log Entry 0====== 00:30:57.031 trtype: tcp 00:30:57.031 adrfam: ipv4 00:30:57.031 subtype: current discovery subsystem 00:30:57.031 treq: not specified, sq flow control disable supported 00:30:57.031 portid: 1 00:30:57.031 trsvcid: 4420 00:30:57.031 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:30:57.031 traddr: 10.0.0.1 00:30:57.031 eflags: none 00:30:57.031 sectype: none 00:30:57.031 =====Discovery Log Entry 1====== 00:30:57.031 trtype: tcp 00:30:57.031 adrfam: ipv4 00:30:57.031 subtype: nvme subsystem 00:30:57.031 treq: not specified, sq flow control disable supported 00:30:57.031 portid: 1 00:30:57.031 trsvcid: 4420 00:30:57.031 subnqn: nqn.2024-02.io.spdk:cnode0 00:30:57.031 traddr: 10.0.0.1 00:30:57.031 eflags: none 00:30:57.031 sectype: none 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@36 -- # mkdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@37 -- # echo 0 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@38 -- # ln -s /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@88 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # IFS=, 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@94 -- # printf %s sha256,sha384,sha512 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # IFS=, 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@94 -- # printf %s ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@93 -- # connect_authenticate sha256,sha384,sha512 ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 1 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256,sha384,sha512 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:57.031 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.032 09:06:46 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.032 nvme0n1 00:30:57.032 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.032 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:57.032 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:57.032 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.032 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.032 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.293 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:57.293 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 0 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 0 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.294 nvme0n1 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:57.294 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 1 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.555 nvme0n1 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:30:57.555 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 2 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.817 nvme0n1 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 3 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 3 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:57.817 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:58.079 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:30:58.079 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.079 09:06:47 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.079 nvme0n1 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 4 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 4 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.079 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.340 nvme0n1 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 0 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 0 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:30:58.340 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.341 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.602 nvme0n1 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 1 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 1 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.602 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.863 nvme0n1 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:58.863 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 2 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 2 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:58.864 09:06:48 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.125 nvme0n1 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 3 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 3 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.125 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.387 nvme0n1 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 4 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 4 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.387 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.649 nvme0n1 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 0 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 0 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:30:59.649 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:30:59.911 09:06:49 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.172 nvme0n1 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 1 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 1 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:00.172 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.173 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.433 nvme0n1 00:31:00.433 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 2 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 2 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.434 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.694 nvme0n1 00:31:00.694 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.694 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:00.694 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:00.694 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.694 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.694 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 3 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 3 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:00.954 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:00.955 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:00.955 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:00.955 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:00.955 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:00.955 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:00.955 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:00.955 09:06:50 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:01.214 nvme0n1 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:01.214 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 4 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 4 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.215 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:01.475 nvme0n1 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 0 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 0 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:01.476 09:06:51 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:02.045 nvme0n1 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:02.045 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 1 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 1 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:02.046 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:02.616 nvme0n1 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 2 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 2 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:02.616 09:06:52 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:03.186 nvme0n1 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 3 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 3 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:03.186 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:03.757 nvme0n1 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 4 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 4 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:03.757 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:03.758 09:06:53 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:04.330 nvme0n1 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 0 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 0 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:04.330 09:06:54 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:05.273 nvme0n1 00:31:05.273 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:05.273 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:05.273 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:05.273 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:05.273 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:05.273 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:05.273 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 1 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 1 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:05.274 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:05.845 nvme0n1 00:31:05.845 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:05.845 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:05.845 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:05.845 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:05.845 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:05.845 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:05.845 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:05.845 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:05.845 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:05.845 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 2 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 2 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:06.106 09:06:55 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:06.678 nvme0n1 00:31:06.678 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:06.678 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:06.678 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:06.678 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:06.678 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:06.678 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:06.678 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:06.678 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:06.678 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:06.678 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 3 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 3 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:06.938 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:06.939 09:06:56 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:07.510 nvme0n1 00:31:07.510 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:07.510 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:07.510 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:07.510 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:07.510 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:07.510 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:07.510 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:07.510 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:07.510 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:07.510 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 4 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 4 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:07.771 09:06:57 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.343 nvme0n1 00:31:08.343 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.343 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:08.343 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:08.343 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.343 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.343 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.343 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:08.343 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:08.343 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 0 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 0 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:08.604 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.605 nvme0n1 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 1 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 1 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:08.605 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.866 nvme0n1 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 2 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 2 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:08.866 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:09.127 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.127 09:06:58 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.127 nvme0n1 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 3 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 3 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.127 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.388 nvme0n1 00:31:09.388 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.388 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:09.388 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:09.388 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.388 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.388 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.388 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 4 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 4 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.389 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.654 nvme0n1 00:31:09.654 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.654 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:09.654 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:09.654 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.654 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.654 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.654 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 0 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 0 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.655 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.917 nvme0n1 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 1 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 1 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:09.917 09:06:59 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.178 nvme0n1 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 2 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 2 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.178 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.439 nvme0n1 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 3 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 3 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.439 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.700 nvme0n1 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 4 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 4 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.700 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.961 nvme0n1 00:31:10.961 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.961 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:10.961 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.961 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:10.961 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.961 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.961 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:10.961 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:10.961 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.961 09:07:00 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 0 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 0 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:10.961 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:11.221 nvme0n1 00:31:11.221 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:11.221 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:11.221 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:11.221 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:11.221 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:11.221 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 1 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 1 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:11.482 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:11.483 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:11.743 nvme0n1 00:31:11.743 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:11.743 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:11.743 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:11.743 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:11.743 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:11.743 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:11.743 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:11.743 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 2 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 2 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:11.744 09:07:01 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.005 nvme0n1 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 3 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 3 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.005 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.265 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.526 nvme0n1 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 4 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 4 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.526 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.787 nvme0n1 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 0 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:12.787 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 0 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:12.788 09:07:02 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:13.358 nvme0n1 00:31:13.358 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.358 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:13.358 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 1 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 1 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.359 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:13.931 nvme0n1 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 2 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 2 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:13.931 09:07:03 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:14.504 nvme0n1 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 3 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:14.504 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 3 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:14.505 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:15.078 nvme0n1 00:31:15.078 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:15.078 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:15.078 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:15.078 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:15.078 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:15.078 09:07:04 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 4 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 4 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:15.078 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:15.650 nvme0n1 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 0 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 0 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:15.650 09:07:05 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:16.592 nvme0n1 00:31:16.592 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 1 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 1 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:16.593 09:07:06 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:17.172 nvme0n1 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 2 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 2 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:17.172 09:07:07 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:18.267 nvme0n1 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 3 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 3 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:18.267 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:18.846 nvme0n1 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 4 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 4 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:18.846 09:07:08 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:19.790 nvme0n1 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 0 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 0 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:19.790 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.052 nvme0n1 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.052 09:07:09 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.052 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:20.052 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 1 00:31:20.052 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:20.052 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:20.052 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:20.052 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:20.052 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 1 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.053 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.315 nvme0n1 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 2 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 2 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.315 nvme0n1 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.315 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 3 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 3 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.577 nvme0n1 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.577 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 4 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 4 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:20.838 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.839 nvme0n1 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 0 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:31:20.839 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 0 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.100 09:07:10 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.100 nvme0n1 00:31:21.100 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.100 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:21.100 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:21.100 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.100 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.100 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.100 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:21.100 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:21.100 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.100 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 1 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 1 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.361 nvme0n1 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.361 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 2 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 2 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:31:21.622 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.623 nvme0n1 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.623 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 3 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 3 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:31:21.884 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.885 nvme0n1 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:21.885 09:07:11 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 4 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 4 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:22.145 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.146 nvme0n1 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.146 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 0 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 0 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.406 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.668 nvme0n1 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 1 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 1 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.668 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.930 nvme0n1 00:31:22.930 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.930 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:22.930 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:22.930 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.930 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.930 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.930 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:22.930 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:22.930 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.930 09:07:12 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 2 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 2 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:22.930 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:23.502 nvme0n1 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 3 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 3 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.502 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:23.763 nvme0n1 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 4 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 4 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:23.763 09:07:13 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:24.024 nvme0n1 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 0 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:24.024 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 0 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:24.025 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:24.595 nvme0n1 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 1 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 1 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:24.595 09:07:14 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:25.165 nvme0n1 00:31:25.165 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:25.165 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:25.165 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:25.165 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:25.165 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:25.165 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:25.165 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 2 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 2 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:25.166 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:25.737 nvme0n1 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 3 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 3 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:25.737 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:25.738 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:25.738 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:25.738 09:07:15 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:26.308 nvme0n1 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 4 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 4 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:26.308 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:26.879 nvme0n1 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 0 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:MWM4NzVkNTBjZDFkOWYyMzRmZDYzZGNmNzhmODgzMTD2rBDr: 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: ]] 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:Y2ExYTYzMDMyOWMwNjNjNDQ5NWZkMzg0MzNhNGFhMGNiNTkxOGUzOGE2NTI2MTY3OGU5NWVhMDMzOTQyOThmMayDVEE=: 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 0 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:26.879 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:31:26.880 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:26.880 09:07:16 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:27.820 nvme0n1 00:31:27.820 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:27.820 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:27.820 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:27.820 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:27.820 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:27.820 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:27.820 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:27.820 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 1 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 1 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:27.821 09:07:17 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:28.392 nvme0n1 00:31:28.392 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:28.392 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:28.392 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:28.392 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:28.392 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:28.392 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:28.392 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:28.392 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:28.392 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:28.392 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:28.652 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:28.652 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:28.652 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 2 00:31:28.652 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:28.652 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:28.652 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 2 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:28.653 09:07:18 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:29.224 nvme0n1 00:31:29.224 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:29.224 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:29.224 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:29.224 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:29.224 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:29.224 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 3 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:YWI5YjEyY2YxY2M2YWEwMmFiMzRmZWEyNTQ5OWQxMzE4ZTYxNjA3MDlkZGUxYzI4Qq/YrA==: 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: ]] 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:MWUxNTczYjdmMjk2ZmI3N2Y2NTVhZGE1OGM2ZjIzMjE+u893: 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 3 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:29.485 09:07:19 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:30.058 nvme0n1 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 4 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:31:30.058 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmYyOTg3M2IyYWJhN2M0NTViYzQwNDg1YzQ3ZmQyZmU4Y2Y4NDA1NTIzNWIyYTAzNThkYzA4YTY5MTNiNDQ4N/Rv2oM=: 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 4 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:30.319 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:30.888 nvme0n1 00:31:30.888 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:30.888 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:31:30.888 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:31:30.888 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:30.888 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:30.888 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:30.888 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:30.888 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:31:30.888 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:30.888 09:07:20 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@111 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@112 -- # get_main_ns_ip 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@112 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:31:31.148 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@650 -- # local es=0 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.149 request: 00:31:31.149 { 00:31:31.149 "name": "nvme0", 00:31:31.149 "trtype": "tcp", 00:31:31.149 "traddr": "10.0.0.1", 00:31:31.149 "adrfam": "ipv4", 00:31:31.149 "trsvcid": "4420", 00:31:31.149 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:31:31.149 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:31:31.149 "prchk_reftag": false, 00:31:31.149 "prchk_guard": false, 00:31:31.149 "hdgst": false, 00:31:31.149 "ddgst": false, 00:31:31.149 "allow_unrecognized_csi": false, 00:31:31.149 "method": "bdev_nvme_attach_controller", 00:31:31.149 "req_id": 1 00:31:31.149 } 00:31:31.149 Got JSON-RPC error response 00:31:31.149 response: 00:31:31.149 { 00:31:31.149 "code": -5, 00:31:31.149 "message": "Input/output error" 00:31:31.149 } 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # es=1 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # rpc_cmd bdev_nvme_get_controllers 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # jq length 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@114 -- # (( 0 == 0 )) 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@117 -- # get_main_ns_ip 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@117 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@650 -- # local es=0 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.149 request: 00:31:31.149 { 00:31:31.149 "name": "nvme0", 00:31:31.149 "trtype": "tcp", 00:31:31.149 "traddr": "10.0.0.1", 00:31:31.149 "adrfam": "ipv4", 00:31:31.149 "trsvcid": "4420", 00:31:31.149 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:31:31.149 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:31:31.149 "prchk_reftag": false, 00:31:31.149 "prchk_guard": false, 00:31:31.149 "hdgst": false, 00:31:31.149 "ddgst": false, 00:31:31.149 "dhchap_key": "key2", 00:31:31.149 "allow_unrecognized_csi": false, 00:31:31.149 "method": "bdev_nvme_attach_controller", 00:31:31.149 "req_id": 1 00:31:31.149 } 00:31:31.149 Got JSON-RPC error response 00:31:31.149 response: 00:31:31.149 { 00:31:31.149 "code": -5, 00:31:31.149 "message": "Input/output error" 00:31:31.149 } 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # es=1 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # rpc_cmd bdev_nvme_get_controllers 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # jq length 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.149 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@120 -- # (( 0 == 0 )) 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@123 -- # get_main_ns_ip 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@123 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@650 -- # local es=0 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.410 request: 00:31:31.410 { 00:31:31.410 "name": "nvme0", 00:31:31.410 "trtype": "tcp", 00:31:31.410 "traddr": "10.0.0.1", 00:31:31.410 "adrfam": "ipv4", 00:31:31.410 "trsvcid": "4420", 00:31:31.410 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:31:31.410 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:31:31.410 "prchk_reftag": false, 00:31:31.410 "prchk_guard": false, 00:31:31.410 "hdgst": false, 00:31:31.410 "ddgst": false, 00:31:31.410 "dhchap_key": "key1", 00:31:31.410 "dhchap_ctrlr_key": "ckey2", 00:31:31.410 "allow_unrecognized_csi": false, 00:31:31.410 "method": "bdev_nvme_attach_controller", 00:31:31.410 "req_id": 1 00:31:31.410 } 00:31:31.410 Got JSON-RPC error response 00:31:31.410 response: 00:31:31.410 { 00:31:31.410 "code": -5, 00:31:31.410 "message": "Input/output error" 00:31:31.410 } 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # es=1 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@128 -- # get_main_ns_ip 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@128 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.410 nvme0n1 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@132 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@133 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.410 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # rpc_cmd bdev_nvme_get_controllers 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # jq -r '.[].name' 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@134 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@136 -- # NOT rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@650 -- # local es=0 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.670 request: 00:31:31.670 { 00:31:31.670 "name": "nvme0", 00:31:31.670 "dhchap_key": "key1", 00:31:31.670 "dhchap_ctrlr_key": "ckey2", 00:31:31.670 "method": "bdev_nvme_set_keys", 00:31:31.670 "req_id": 1 00:31:31.670 } 00:31:31.670 Got JSON-RPC error response 00:31:31.670 response: 00:31:31.670 { 00:31:31.670 "code": -13, 00:31:31.670 "message": "Permission denied" 00:31:31.670 } 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # es=1 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 1 != 0 )) 00:31:31.670 09:07:21 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@138 -- # sleep 1s 00:31:33.053 09:07:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:31:33.053 09:07:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:31:33.053 09:07:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:33.053 09:07:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:33.053 09:07:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:33.053 09:07:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 1 != 0 )) 00:31:33.053 09:07:22 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@138 -- # sleep 1s 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # rpc_cmd bdev_nvme_get_controllers 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # jq length 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@137 -- # (( 0 != 0 )) 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@141 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:OTgxYTNjYzM5OGFhZjEzYzZkMzAwOGNiMzk3NTJmOGVhNzNhNzgyZGJmMjI5NWE4eB+WxA==: 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: ]] 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:OGIwMDQ5YzMyMTM1ZDcxMmZiZmVkMmQyMWVkMWQwYjA5MTc5NWZlZDBmYTkzNDYwv5llug==: 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@142 -- # get_main_ns_ip 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@767 -- # local ip 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # ip_candidates=() 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@768 -- # local -A ip_candidates 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@142 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 --ctrlr-loss-timeout-sec 1 --reconnect-delay-sec 1 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:33.995 nvme0n1 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@146 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:ZmQ1NWJmZjdkNTIwZGJiYjk1Zjk0MTE4YjYwMzkwNTInSzPO: 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: ]] 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NjZhOTQ1MTJiMjVkYjcwM2IyNThhNTU4Nzk4NGEyNDG5Tlby: 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@147 -- # NOT rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@650 -- # local es=0 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # rpc_cmd bdev_nvme_set_keys nvme0 --dhchap-key key2 --dhchap-ctrlr-key ckey1 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:33.995 09:07:23 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:33.995 request: 00:31:33.995 { 00:31:33.995 "name": "nvme0", 00:31:33.995 "dhchap_key": "key2", 00:31:33.995 "dhchap_ctrlr_key": "ckey1", 00:31:33.995 "method": "bdev_nvme_set_keys", 00:31:33.995 "req_id": 1 00:31:33.995 } 00:31:33.995 Got JSON-RPC error response 00:31:33.995 response: 00:31:33.995 { 00:31:33.995 "code": -13, 00:31:33.995 "message": "Permission denied" 00:31:33.995 } 00:31:33.995 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:31:33.995 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@653 -- # es=1 00:31:33.995 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:31:33.995 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:31:33.995 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:31:33.995 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # rpc_cmd bdev_nvme_get_controllers 00:31:33.995 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # jq length 00:31:33.995 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:33.995 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:33.995 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:34.256 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # (( 1 != 0 )) 00:31:34.256 09:07:24 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@149 -- # sleep 1s 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # rpc_cmd bdev_nvme_get_controllers 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # jq length 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@148 -- # (( 0 != 0 )) 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@152 -- # trap - SIGINT SIGTERM EXIT 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@153 -- # cleanup 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@24 -- # nvmftestfini 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@514 -- # nvmfcleanup 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@121 -- # sync 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@124 -- # set +e 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@125 -- # for i in {1..20} 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:31:35.197 rmmod nvme_tcp 00:31:35.197 rmmod nvme_fabrics 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@128 -- # set -e 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@129 -- # return 0 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@515 -- # '[' -n 2044323 ']' 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@516 -- # killprocess 2044323 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@950 -- # '[' -z 2044323 ']' 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@954 -- # kill -0 2044323 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@955 -- # uname 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2044323 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2044323' 00:31:35.197 killing process with pid 2044323 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@969 -- # kill 2044323 00:31:35.197 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@974 -- # wait 2044323 00:31:35.457 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:31:35.457 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:31:35.457 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:31:35.457 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@297 -- # iptr 00:31:35.457 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@789 -- # iptables-save 00:31:35.458 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:31:35.458 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@789 -- # iptables-restore 00:31:35.458 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:31:35.458 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@302 -- # remove_spdk_ns 00:31:35.458 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:35.458 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:35.458 09:07:25 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:37.370 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:31:37.370 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@25 -- # rm /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@26 -- # rmdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@27 -- # clean_kernel_target 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@710 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 ]] 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@712 -- # echo 0 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@714 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2024-02.io.spdk:cnode0 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@715 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@716 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@719 -- # modules=(/sys/module/nvmet/holders/*) 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@721 -- # modprobe -r nvmet_tcp nvmet 00:31:37.632 09:07:27 nvmf_tcp.nvmf_host.nvmf_auth_host -- nvmf/common.sh@724 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:31:40.934 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:31:40.934 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:31:40.934 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:31:40.934 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:31:40.934 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:31:40.934 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:31:40.934 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:31:41.195 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:31:41.195 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:31:41.195 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:31:41.195 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:31:41.195 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:31:41.195 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:31:41.195 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:31:41.195 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:31:41.195 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:31:41.195 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:31:41.456 09:07:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@28 -- # rm -f /tmp/spdk.key-null.s1J /tmp/spdk.key-null.Igb /tmp/spdk.key-sha256.knh /tmp/spdk.key-sha384.bJm /tmp/spdk.key-sha512.AYV /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log 00:31:41.456 09:07:31 nvmf_tcp.nvmf_host.nvmf_auth_host -- host/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:31:45.664 0000:80:01.6 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:80:01.7 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:80:01.4 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:80:01.5 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:80:01.2 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:80:01.3 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:80:01.0 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:80:01.1 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:00:01.6 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:65:00.0 (144d a80a): Already using the vfio-pci driver 00:31:45.664 0000:00:01.7 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:00:01.4 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:00:01.5 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:00:01.2 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:00:01.3 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:00:01.0 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 0000:00:01.1 (8086 0b00): Already using the vfio-pci driver 00:31:45.664 00:31:45.664 real 1m3.424s 00:31:45.664 user 0m57.150s 00:31:45.664 sys 0m15.860s 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@1126 -- # xtrace_disable 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:31:45.664 ************************************ 00:31:45.664 END TEST nvmf_auth_host 00:31:45.664 ************************************ 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@32 -- # [[ tcp == \t\c\p ]] 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@33 -- # run_test nvmf_digest /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:31:45.664 ************************************ 00:31:45.664 START TEST nvmf_digest 00:31:45.664 ************************************ 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:31:45.664 * Looking for test storage... 00:31:45.664 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1689 -- # lcov --version 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:31:45.664 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@333 -- # local ver1 ver1_l 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@334 -- # local ver2 ver2_l 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@336 -- # IFS=.-: 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@336 -- # read -ra ver1 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@337 -- # IFS=.-: 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@337 -- # read -ra ver2 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@338 -- # local 'op=<' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@340 -- # ver1_l=2 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@341 -- # ver2_l=1 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@344 -- # case "$op" in 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@345 -- # : 1 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@364 -- # (( v = 0 )) 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@365 -- # decimal 1 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@353 -- # local d=1 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@355 -- # echo 1 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@365 -- # ver1[v]=1 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@366 -- # decimal 2 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@353 -- # local d=2 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@355 -- # echo 2 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@366 -- # ver2[v]=2 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@368 -- # return 0 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:31:45.665 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:45.665 --rc genhtml_branch_coverage=1 00:31:45.665 --rc genhtml_function_coverage=1 00:31:45.665 --rc genhtml_legend=1 00:31:45.665 --rc geninfo_all_blocks=1 00:31:45.665 --rc geninfo_unexecuted_blocks=1 00:31:45.665 00:31:45.665 ' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:31:45.665 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:45.665 --rc genhtml_branch_coverage=1 00:31:45.665 --rc genhtml_function_coverage=1 00:31:45.665 --rc genhtml_legend=1 00:31:45.665 --rc geninfo_all_blocks=1 00:31:45.665 --rc geninfo_unexecuted_blocks=1 00:31:45.665 00:31:45.665 ' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:31:45.665 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:45.665 --rc genhtml_branch_coverage=1 00:31:45.665 --rc genhtml_function_coverage=1 00:31:45.665 --rc genhtml_legend=1 00:31:45.665 --rc geninfo_all_blocks=1 00:31:45.665 --rc geninfo_unexecuted_blocks=1 00:31:45.665 00:31:45.665 ' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:31:45.665 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:31:45.665 --rc genhtml_branch_coverage=1 00:31:45.665 --rc genhtml_function_coverage=1 00:31:45.665 --rc genhtml_legend=1 00:31:45.665 --rc geninfo_all_blocks=1 00:31:45.665 --rc geninfo_unexecuted_blocks=1 00:31:45.665 00:31:45.665 ' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@7 -- # uname -s 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@15 -- # shopt -s extglob 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@5 -- # export PATH 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@51 -- # : 0 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:31:45.665 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@55 -- # have_pci_nics=0 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@14 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@15 -- # bperfsock=/var/tmp/bperf.sock 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@16 -- # runtime=2 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@136 -- # [[ tcp != \t\c\p ]] 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@138 -- # nvmftestinit 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@474 -- # prepare_net_devs 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@436 -- # local -g is_hw=no 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@438 -- # remove_spdk_ns 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@309 -- # xtrace_disable 00:31:45.665 09:07:35 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@315 -- # pci_devs=() 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@315 -- # local -a pci_devs 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@316 -- # pci_net_devs=() 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@317 -- # pci_drivers=() 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@317 -- # local -A pci_drivers 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@319 -- # net_devs=() 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@319 -- # local -ga net_devs 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@320 -- # e810=() 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@320 -- # local -ga e810 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@321 -- # x722=() 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@321 -- # local -ga x722 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@322 -- # mlx=() 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@322 -- # local -ga mlx 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:31:53.807 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:31:53.807 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@416 -- # [[ up == up ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:31:53.807 Found net devices under 0000:4b:00.0: cvl_0_0 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@416 -- # [[ up == up ]] 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:31:53.807 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:31:53.808 Found net devices under 0000:4b:00.1: cvl_0_1 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@440 -- # is_hw=yes 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:31:53.808 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:31:53.808 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.600 ms 00:31:53.808 00:31:53.808 --- 10.0.0.2 ping statistics --- 00:31:53.808 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:53.808 rtt min/avg/max/mdev = 0.600/0.600/0.600/0.000 ms 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:31:53.808 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:31:53.808 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.278 ms 00:31:53.808 00:31:53.808 --- 10.0.0.1 ping statistics --- 00:31:53.808 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:31:53.808 rtt min/avg/max/mdev = 0.278/0.278/0.278/0.000 ms 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@448 -- # return 0 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:31:53.808 09:07:42 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@140 -- # trap cleanup SIGINT SIGTERM EXIT 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@141 -- # [[ 0 -eq 1 ]] 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@145 -- # run_test nvmf_digest_clean run_digest 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1107 -- # xtrace_disable 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:31:53.808 ************************************ 00:31:53.808 START TEST nvmf_digest_clean 00:31:53.808 ************************************ 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@1125 -- # run_digest 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@120 -- # local dsa_initiator 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@121 -- # [[ '' == \d\s\a\_\i\n\i\t\i\a\t\o\r ]] 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@121 -- # dsa_initiator=false 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@123 -- # tgt_params=("--wait-for-rpc") 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@124 -- # nvmfappstart --wait-for-rpc 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@724 -- # xtrace_disable 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@507 -- # nvmfpid=2061940 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@508 -- # waitforlisten 2061940 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@831 -- # '[' -z 2061940 ']' 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:31:53.808 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:31:53.808 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:31:53.809 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:31:53.809 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:31:53.809 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:31:53.809 [2024-11-06 09:07:43.121679] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:31:53.809 [2024-11-06 09:07:43.121739] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:31:53.809 [2024-11-06 09:07:43.206424] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:53.809 [2024-11-06 09:07:43.246616] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:31:53.809 [2024-11-06 09:07:43.246656] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:31:53.809 [2024-11-06 09:07:43.246664] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:31:53.809 [2024-11-06 09:07:43.246671] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:31:53.809 [2024-11-06 09:07:43.246677] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:31:53.809 [2024-11-06 09:07:43.247314] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:31:53.809 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:31:53.809 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # return 0 00:31:53.809 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:31:53.809 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@730 -- # xtrace_disable 00:31:53.809 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:31:54.069 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:31:54.069 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@125 -- # [[ '' == \d\s\a\_\t\a\r\g\e\t ]] 00:31:54.069 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@126 -- # common_target_config 00:31:54.069 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@43 -- # rpc_cmd 00:31:54.069 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@561 -- # xtrace_disable 00:31:54.069 09:07:43 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:31:54.069 null0 00:31:54.069 [2024-11-06 09:07:44.024854] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:31:54.069 [2024-11-06 09:07:44.049061] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:31:54.069 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:31:54.069 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@128 -- # run_bperf randread 4096 128 false 00:31:54.069 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:31:54.069 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:31:54.069 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randread 00:31:54.069 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=4096 00:31:54.069 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=128 00:31:54.069 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:31:54.070 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=2062013 00:31:54.070 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 2062013 /var/tmp/bperf.sock 00:31:54.070 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@831 -- # '[' -z 2062013 ']' 00:31:54.070 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:31:54.070 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:31:54.070 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:31:54.070 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:31:54.070 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:31:54.070 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:31:54.070 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:31:54.070 [2024-11-06 09:07:44.103169] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:31:54.070 [2024-11-06 09:07:44.103219] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2062013 ] 00:31:54.330 [2024-11-06 09:07:44.190770] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:54.330 [2024-11-06 09:07:44.226745] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:54.901 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:31:54.901 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # return 0 00:31:54.901 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:31:54.901 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:31:54.901 09:07:44 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:31:55.161 09:07:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:31:55.161 09:07:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:31:55.422 nvme0n1 00:31:55.422 09:07:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:31:55.422 09:07:45 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:31:55.682 Running I/O for 2 seconds... 00:31:57.565 19283.00 IOPS, 75.32 MiB/s [2024-11-06T08:07:47.678Z] 19303.50 IOPS, 75.40 MiB/s 00:31:57.565 Latency(us) 00:31:57.565 [2024-11-06T08:07:47.678Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:57.565 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:31:57.565 nvme0n1 : 2.00 19320.61 75.47 0.00 0.00 6618.48 2839.89 16165.55 00:31:57.565 [2024-11-06T08:07:47.678Z] =================================================================================================================== 00:31:57.565 [2024-11-06T08:07:47.678Z] Total : 19320.61 75.47 0.00 0.00 6618.48 2839.89 16165.55 00:31:57.565 { 00:31:57.565 "results": [ 00:31:57.565 { 00:31:57.565 "job": "nvme0n1", 00:31:57.565 "core_mask": "0x2", 00:31:57.565 "workload": "randread", 00:31:57.565 "status": "finished", 00:31:57.565 "queue_depth": 128, 00:31:57.565 "io_size": 4096, 00:31:57.565 "runtime": 2.004854, 00:31:57.565 "iops": 19320.608882242796, 00:31:57.565 "mibps": 75.47112844626092, 00:31:57.565 "io_failed": 0, 00:31:57.565 "io_timeout": 0, 00:31:57.565 "avg_latency_us": 6618.480562454284, 00:31:57.565 "min_latency_us": 2839.8933333333334, 00:31:57.565 "max_latency_us": 16165.546666666667 00:31:57.565 } 00:31:57.565 ], 00:31:57.565 "core_count": 1 00:31:57.565 } 00:31:57.565 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:31:57.565 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:31:57.565 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:31:57.565 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:31:57.565 | select(.opcode=="crc32c") 00:31:57.565 | "\(.module_name) \(.executed)"' 00:31:57.565 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 2062013 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@950 -- # '[' -z 2062013 ']' 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # kill -0 2062013 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # uname 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2062013 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2062013' 00:31:57.831 killing process with pid 2062013 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@969 -- # kill 2062013 00:31:57.831 Received shutdown signal, test time was about 2.000000 seconds 00:31:57.831 00:31:57.831 Latency(us) 00:31:57.831 [2024-11-06T08:07:47.944Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:31:57.831 [2024-11-06T08:07:47.944Z] =================================================================================================================== 00:31:57.831 [2024-11-06T08:07:47.944Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:31:57.831 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@974 -- # wait 2062013 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@129 -- # run_bperf randread 131072 16 false 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randread 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=131072 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=16 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=2062808 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 2062808 /var/tmp/bperf.sock 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@831 -- # '[' -z 2062808 ']' 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:31:58.127 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:31:58.127 09:07:47 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:31:58.127 [2024-11-06 09:07:48.017109] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:31:58.127 [2024-11-06 09:07:48.017166] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2062808 ] 00:31:58.127 I/O size of 131072 is greater than zero copy threshold (65536). 00:31:58.127 Zero copy mechanism will not be used. 00:31:58.127 [2024-11-06 09:07:48.098705] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:31:58.127 [2024-11-06 09:07:48.128118] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:31:58.718 09:07:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:31:58.718 09:07:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # return 0 00:31:58.719 09:07:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:31:58.719 09:07:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:31:58.719 09:07:48 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:31:58.979 09:07:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:31:58.979 09:07:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:31:59.549 nvme0n1 00:31:59.549 09:07:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:31:59.549 09:07:49 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:31:59.549 I/O size of 131072 is greater than zero copy threshold (65536). 00:31:59.549 Zero copy mechanism will not be used. 00:31:59.549 Running I/O for 2 seconds... 00:32:01.429 4004.00 IOPS, 500.50 MiB/s [2024-11-06T08:07:51.542Z] 4080.50 IOPS, 510.06 MiB/s 00:32:01.429 Latency(us) 00:32:01.429 [2024-11-06T08:07:51.542Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:01.429 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:32:01.429 nvme0n1 : 2.00 4081.95 510.24 0.00 0.00 3916.58 1133.23 13817.17 00:32:01.429 [2024-11-06T08:07:51.542Z] =================================================================================================================== 00:32:01.429 [2024-11-06T08:07:51.542Z] Total : 4081.95 510.24 0.00 0.00 3916.58 1133.23 13817.17 00:32:01.429 { 00:32:01.429 "results": [ 00:32:01.429 { 00:32:01.429 "job": "nvme0n1", 00:32:01.429 "core_mask": "0x2", 00:32:01.429 "workload": "randread", 00:32:01.429 "status": "finished", 00:32:01.429 "queue_depth": 16, 00:32:01.429 "io_size": 131072, 00:32:01.429 "runtime": 2.003209, 00:32:01.429 "iops": 4081.9505104060536, 00:32:01.429 "mibps": 510.2438138007567, 00:32:01.429 "io_failed": 0, 00:32:01.429 "io_timeout": 0, 00:32:01.429 "avg_latency_us": 3916.579805144511, 00:32:01.429 "min_latency_us": 1133.2266666666667, 00:32:01.429 "max_latency_us": 13817.173333333334 00:32:01.429 } 00:32:01.429 ], 00:32:01.429 "core_count": 1 00:32:01.429 } 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:32:01.690 | select(.opcode=="crc32c") 00:32:01.690 | "\(.module_name) \(.executed)"' 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 2062808 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@950 -- # '[' -z 2062808 ']' 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # kill -0 2062808 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # uname 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2062808 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2062808' 00:32:01.690 killing process with pid 2062808 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@969 -- # kill 2062808 00:32:01.690 Received shutdown signal, test time was about 2.000000 seconds 00:32:01.690 00:32:01.690 Latency(us) 00:32:01.690 [2024-11-06T08:07:51.803Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:01.690 [2024-11-06T08:07:51.803Z] =================================================================================================================== 00:32:01.690 [2024-11-06T08:07:51.803Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:01.690 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@974 -- # wait 2062808 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@130 -- # run_bperf randwrite 4096 128 false 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randwrite 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=4096 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=128 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=2063662 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 2063662 /var/tmp/bperf.sock 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@831 -- # '[' -z 2063662 ']' 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:32:01.950 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:32:01.950 09:07:51 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:32:01.950 [2024-11-06 09:07:51.934673] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:01.951 [2024-11-06 09:07:51.934733] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2063662 ] 00:32:01.951 [2024-11-06 09:07:52.016935] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:01.951 [2024-11-06 09:07:52.046173] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:02.891 09:07:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:02.891 09:07:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # return 0 00:32:02.891 09:07:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:32:02.891 09:07:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:32:02.891 09:07:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:32:02.891 09:07:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:02.891 09:07:52 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:03.151 nvme0n1 00:32:03.151 09:07:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:32:03.151 09:07:53 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:32:03.412 Running I/O for 2 seconds... 00:32:05.295 21531.00 IOPS, 84.11 MiB/s [2024-11-06T08:07:55.408Z] 21606.00 IOPS, 84.40 MiB/s 00:32:05.295 Latency(us) 00:32:05.295 [2024-11-06T08:07:55.408Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:05.295 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:32:05.295 nvme0n1 : 2.01 21626.40 84.48 0.00 0.00 5910.29 2020.69 11796.48 00:32:05.295 [2024-11-06T08:07:55.408Z] =================================================================================================================== 00:32:05.295 [2024-11-06T08:07:55.408Z] Total : 21626.40 84.48 0.00 0.00 5910.29 2020.69 11796.48 00:32:05.295 { 00:32:05.295 "results": [ 00:32:05.295 { 00:32:05.295 "job": "nvme0n1", 00:32:05.295 "core_mask": "0x2", 00:32:05.295 "workload": "randwrite", 00:32:05.295 "status": "finished", 00:32:05.295 "queue_depth": 128, 00:32:05.295 "io_size": 4096, 00:32:05.295 "runtime": 2.006252, 00:32:05.295 "iops": 21626.39588645893, 00:32:05.295 "mibps": 84.4781089314802, 00:32:05.295 "io_failed": 0, 00:32:05.295 "io_timeout": 0, 00:32:05.295 "avg_latency_us": 5910.286420208352, 00:32:05.295 "min_latency_us": 2020.6933333333334, 00:32:05.295 "max_latency_us": 11796.48 00:32:05.295 } 00:32:05.295 ], 00:32:05.295 "core_count": 1 00:32:05.295 } 00:32:05.295 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:32:05.295 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:32:05.295 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:32:05.295 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:32:05.295 | select(.opcode=="crc32c") 00:32:05.295 | "\(.module_name) \(.executed)"' 00:32:05.295 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 2063662 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@950 -- # '[' -z 2063662 ']' 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # kill -0 2063662 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # uname 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2063662 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2063662' 00:32:05.557 killing process with pid 2063662 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@969 -- # kill 2063662 00:32:05.557 Received shutdown signal, test time was about 2.000000 seconds 00:32:05.557 00:32:05.557 Latency(us) 00:32:05.557 [2024-11-06T08:07:55.670Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:05.557 [2024-11-06T08:07:55.670Z] =================================================================================================================== 00:32:05.557 [2024-11-06T08:07:55.670Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:05.557 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@974 -- # wait 2063662 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@131 -- # run_bperf randwrite 131072 16 false 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randwrite 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=131072 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=16 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=2064409 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 2064409 /var/tmp/bperf.sock 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@831 -- # '[' -z 2064409 ']' 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:32:05.818 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:32:05.818 [2024-11-06 09:07:55.738766] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:05.818 [2024-11-06 09:07:55.738822] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2064409 ] 00:32:05.818 I/O size of 131072 is greater than zero copy threshold (65536). 00:32:05.818 Zero copy mechanism will not be used. 00:32:05.818 [2024-11-06 09:07:55.822557] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:05.818 [2024-11-06 09:07:55.851635] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@864 -- # return 0 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:32:05.818 09:07:55 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:32:06.079 09:07:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:06.079 09:07:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:06.340 nvme0n1 00:32:06.340 09:07:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:32:06.340 09:07:56 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:32:06.340 I/O size of 131072 is greater than zero copy threshold (65536). 00:32:06.340 Zero copy mechanism will not be used. 00:32:06.340 Running I/O for 2 seconds... 00:32:08.667 6185.00 IOPS, 773.12 MiB/s [2024-11-06T08:07:58.780Z] 6053.50 IOPS, 756.69 MiB/s 00:32:08.667 Latency(us) 00:32:08.667 [2024-11-06T08:07:58.781Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:08.668 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:32:08.668 nvme0n1 : 2.01 6043.83 755.48 0.00 0.00 2641.54 1433.60 11578.03 00:32:08.668 [2024-11-06T08:07:58.781Z] =================================================================================================================== 00:32:08.668 [2024-11-06T08:07:58.781Z] Total : 6043.83 755.48 0.00 0.00 2641.54 1433.60 11578.03 00:32:08.668 { 00:32:08.668 "results": [ 00:32:08.668 { 00:32:08.668 "job": "nvme0n1", 00:32:08.668 "core_mask": "0x2", 00:32:08.668 "workload": "randwrite", 00:32:08.668 "status": "finished", 00:32:08.668 "queue_depth": 16, 00:32:08.668 "io_size": 131072, 00:32:08.668 "runtime": 2.006344, 00:32:08.668 "iops": 6043.828974492909, 00:32:08.668 "mibps": 755.4786218116136, 00:32:08.668 "io_failed": 0, 00:32:08.668 "io_timeout": 0, 00:32:08.668 "avg_latency_us": 2641.539145637473, 00:32:08.668 "min_latency_us": 1433.6, 00:32:08.668 "max_latency_us": 11578.026666666667 00:32:08.668 } 00:32:08.668 ], 00:32:08.668 "core_count": 1 00:32:08.668 } 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:32:08.668 | select(.opcode=="crc32c") 00:32:08.668 | "\(.module_name) \(.executed)"' 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 2064409 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@950 -- # '[' -z 2064409 ']' 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # kill -0 2064409 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # uname 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2064409 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2064409' 00:32:08.668 killing process with pid 2064409 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@969 -- # kill 2064409 00:32:08.668 Received shutdown signal, test time was about 2.000000 seconds 00:32:08.668 00:32:08.668 Latency(us) 00:32:08.668 [2024-11-06T08:07:58.781Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:08.668 [2024-11-06T08:07:58.781Z] =================================================================================================================== 00:32:08.668 [2024-11-06T08:07:58.781Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@974 -- # wait 2064409 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- host/digest.sh@132 -- # killprocess 2061940 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@950 -- # '[' -z 2061940 ']' 00:32:08.668 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # kill -0 2061940 00:32:08.929 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # uname 00:32:08.929 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:08.929 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2061940 00:32:08.929 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:32:08.929 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:32:08.929 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2061940' 00:32:08.929 killing process with pid 2061940 00:32:08.929 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@969 -- # kill 2061940 00:32:08.929 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@974 -- # wait 2061940 00:32:08.929 00:32:08.929 real 0m15.900s 00:32:08.929 user 0m31.380s 00:32:08.929 sys 0m3.440s 00:32:08.929 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:08.929 09:07:58 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:32:08.929 ************************************ 00:32:08.929 END TEST nvmf_digest_clean 00:32:08.929 ************************************ 00:32:08.929 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@147 -- # run_test nvmf_digest_error run_digest_error 00:32:08.929 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:32:08.929 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1107 -- # xtrace_disable 00:32:08.929 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:32:08.929 ************************************ 00:32:08.929 START TEST nvmf_digest_error 00:32:08.929 ************************************ 00:32:08.929 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@1125 -- # run_digest_error 00:32:08.929 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@102 -- # nvmfappstart --wait-for-rpc 00:32:08.929 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:32:08.929 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@724 -- # xtrace_disable 00:32:08.929 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:09.190 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@507 -- # nvmfpid=2065108 00:32:09.190 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@508 -- # waitforlisten 2065108 00:32:09.190 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@831 -- # '[' -z 2065108 ']' 00:32:09.191 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:09.191 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:09.191 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:09.191 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:09.191 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:09.191 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:09.191 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:32:09.191 [2024-11-06 09:07:59.095213] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:09.191 [2024-11-06 09:07:59.095263] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:09.191 [2024-11-06 09:07:59.172559] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:09.191 [2024-11-06 09:07:59.208629] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:09.191 [2024-11-06 09:07:59.208662] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:09.191 [2024-11-06 09:07:59.208675] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:09.191 [2024-11-06 09:07:59.208681] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:09.191 [2024-11-06 09:07:59.208687] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:09.191 [2024-11-06 09:07:59.209255] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # return 0 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@730 -- # xtrace_disable 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@104 -- # rpc_cmd accel_assign_opc -o crc32c -m error 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:10.132 [2024-11-06 09:07:59.923307] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation crc32c will be assigned to module error 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@105 -- # common_target_config 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@43 -- # rpc_cmd 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:10.132 09:07:59 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:10.132 null0 00:32:10.132 [2024-11-06 09:08:00.002619] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:10.133 [2024-11-06 09:08:00.026827] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@108 -- # run_bperf_err randread 4096 128 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randread 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=4096 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=128 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=2065151 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 2065151 /var/tmp/bperf.sock 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@831 -- # '[' -z 2065151 ']' 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:32:10.133 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:10.133 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z 00:32:10.133 [2024-11-06 09:08:00.081624] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:10.133 [2024-11-06 09:08:00.081673] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2065151 ] 00:32:10.133 [2024-11-06 09:08:00.164210] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:10.133 [2024-11-06 09:08:00.193867] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:11.075 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:11.075 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # return 0 00:32:11.075 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:32:11.075 09:08:00 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:32:11.075 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:32:11.075 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:11.075 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:11.075 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:11.075 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:11.075 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:11.336 nvme0n1 00:32:11.336 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:32:11.336 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:11.336 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:11.597 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:11.597 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:32:11.597 09:08:01 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:32:11.597 Running I/O for 2 seconds... 00:32:11.597 [2024-11-06 09:08:01.564950] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.564981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:15350 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.564991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.576624] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.576643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:17044 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.576651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.589216] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.589234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:1128 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.589240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.603948] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.603966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:9841 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.603973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.613777] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.613794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:20090 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.613801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.626716] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.626733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:2314 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.626740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.641455] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.641472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:20474 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.641478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.654572] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.654590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:20873 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.654596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.666061] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.666079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:19135 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.666085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.677702] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.677719] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:3866 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.677726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.690918] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.690936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:13907 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.690949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.597 [2024-11-06 09:08:01.703900] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.597 [2024-11-06 09:08:01.703918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:18455 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.597 [2024-11-06 09:08:01.703924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.714751] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.714768] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:8362 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.714776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.726470] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.726487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:15415 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.726494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.739960] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.739979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:1486 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.739986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.753101] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.753119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:24004 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.753125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.768049] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.768066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:7823 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.768073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.779248] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.779265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:2499 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.779271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.791125] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.791142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:9784 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.791149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.804872] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.804893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:9002 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.804899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.816937] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.816954] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:10039 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.816961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.827603] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.827620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:5391 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.827627] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.840103] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.840119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:17104 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.840126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.853808] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.853825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:18388 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.853832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:100 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.868259] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.868276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:6100 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.868282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.877518] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.877534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:22876 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.858 [2024-11-06 09:08:01.877540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.858 [2024-11-06 09:08:01.892386] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.858 [2024-11-06 09:08:01.892403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:3141 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.859 [2024-11-06 09:08:01.892409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.859 [2024-11-06 09:08:01.903813] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.859 [2024-11-06 09:08:01.903830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:5177 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.859 [2024-11-06 09:08:01.903837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.859 [2024-11-06 09:08:01.916741] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.859 [2024-11-06 09:08:01.916763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:9742 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.859 [2024-11-06 09:08:01.916769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.859 [2024-11-06 09:08:01.929348] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.859 [2024-11-06 09:08:01.929365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:13788 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.859 [2024-11-06 09:08:01.929372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.859 [2024-11-06 09:08:01.941949] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.859 [2024-11-06 09:08:01.941966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:14001 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.859 [2024-11-06 09:08:01.941972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.859 [2024-11-06 09:08:01.955282] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.859 [2024-11-06 09:08:01.955299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:9156 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.859 [2024-11-06 09:08:01.955306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:11.859 [2024-11-06 09:08:01.967418] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:11.859 [2024-11-06 09:08:01.967435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:10745 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:11.859 [2024-11-06 09:08:01.967442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:47 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:01.980473] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:01.980490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:23472 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:01.980497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:89 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:01.991901] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:01.991918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:21412 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:01.991925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.005742] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.005763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:22221 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.005769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.017671] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.017688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:20506 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.017698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:94 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.029328] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.029345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:2025 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.029352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.042123] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.042141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:4123 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.042147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.055856] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.055874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:16544 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.055880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.068825] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.068842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:24425 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.068849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.080538] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.080555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:10635 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.080562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.093534] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.093551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:10883 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.093557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.105539] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.105556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:14708 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.105563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:112 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.117459] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.117477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:2890 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.117483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.130118] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.130135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:24453 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.130142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.143837] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.143854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:6330 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.143861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.157210] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.157227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:4212 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.157234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.167083] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.167100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:24978 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.167107] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.180804] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.120 [2024-11-06 09:08:02.180821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:25273 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.120 [2024-11-06 09:08:02.180828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:112 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.120 [2024-11-06 09:08:02.193059] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.121 [2024-11-06 09:08:02.193076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:24076 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.121 [2024-11-06 09:08:02.193083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.121 [2024-11-06 09:08:02.206377] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.121 [2024-11-06 09:08:02.206394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:13754 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.121 [2024-11-06 09:08:02.206400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.121 [2024-11-06 09:08:02.219388] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.121 [2024-11-06 09:08:02.219404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:22663 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.121 [2024-11-06 09:08:02.219411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.121 [2024-11-06 09:08:02.231207] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.121 [2024-11-06 09:08:02.231224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:6138 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.121 [2024-11-06 09:08:02.231234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.243080] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.243098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:13625 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.243104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.256377] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.256395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:17486 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.256402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.268589] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.268607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:12510 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.268614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.281475] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.281492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:21000 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.281499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.294097] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.294114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:3692 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.294121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.306511] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.306528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:8234 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.306535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.319639] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.319656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:15293 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.319662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.332612] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.332629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:7061 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.332636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:57 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.345029] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.345050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:325 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.345056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:100 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.358263] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.358280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:21560 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.358286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.368662] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.368679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:5307 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.368685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.383166] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.383183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:20802 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.383189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.395129] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.395146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:16080 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.395152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.407866] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.407882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:16464 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.407889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.420619] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.420635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:6106 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.420641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.431756] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.431774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:16289 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.431780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.444630] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.444647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:7050 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.382 [2024-11-06 09:08:02.444654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.382 [2024-11-06 09:08:02.457399] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.382 [2024-11-06 09:08:02.457416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:16190 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.383 [2024-11-06 09:08:02.457422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.383 [2024-11-06 09:08:02.470597] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.383 [2024-11-06 09:08:02.470613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:21395 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.383 [2024-11-06 09:08:02.470620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.383 [2024-11-06 09:08:02.482737] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.383 [2024-11-06 09:08:02.482756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:19763 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.383 [2024-11-06 09:08:02.482763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:57 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.644 [2024-11-06 09:08:02.495152] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.644 [2024-11-06 09:08:02.495170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:22881 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.644 [2024-11-06 09:08:02.495177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.644 [2024-11-06 09:08:02.510132] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.644 [2024-11-06 09:08:02.510149] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:18617 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.644 [2024-11-06 09:08:02.510156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.644 [2024-11-06 09:08:02.523134] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.644 [2024-11-06 09:08:02.523151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:3785 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.644 [2024-11-06 09:08:02.523157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:120 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.644 [2024-11-06 09:08:02.534855] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.644 [2024-11-06 09:08:02.534872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:7274 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.644 [2024-11-06 09:08:02.534878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.644 19991.00 IOPS, 78.09 MiB/s [2024-11-06T08:08:02.757Z] [2024-11-06 09:08:02.549357] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.644 [2024-11-06 09:08:02.549373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:11700 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.644 [2024-11-06 09:08:02.549380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.644 [2024-11-06 09:08:02.561836] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.644 [2024-11-06 09:08:02.561853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:13165 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.644 [2024-11-06 09:08:02.561864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.644 [2024-11-06 09:08:02.572835] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.644 [2024-11-06 09:08:02.572852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:723 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.644 [2024-11-06 09:08:02.572858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.644 [2024-11-06 09:08:02.586680] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.586697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:10664 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.586703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.599307] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.599324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:9038 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.599331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.610414] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.610431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:1776 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.610437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.624382] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.624399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:9590 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.624406] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.637450] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.637467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:18748 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.637475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.647365] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.647382] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:7279 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.647388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.660314] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.660331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:15824 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.660338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.674260] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.674279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:13031 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.674286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.687147] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.687164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:12651 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.687170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.698324] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.698340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:22722 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.698346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.711498] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.711514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:1922 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.711520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:47 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.723327] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.723343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:10189 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.723349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:119 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.736005] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.736022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:17509 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.736029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.645 [2024-11-06 09:08:02.749231] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.645 [2024-11-06 09:08:02.749248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:5991 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.645 [2024-11-06 09:08:02.749254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.762909] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.762926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:15987 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.762933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.774676] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.774693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:22331 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.774699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.786121] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.786138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:3659 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.786144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.799119] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.799136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:234 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.799142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.811999] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.812015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:14843 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.812022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.823501] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.823517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:5305 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.823524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.836761] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.836778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:9007 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.836784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.849462] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.849478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:23880 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.849485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.862185] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.862200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:7266 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.862207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.875488] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.875504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:1596 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.875511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.888022] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.888041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:3345 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.888048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.897755] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.897771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:14832 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.897778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.910678] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.910694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:3301 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.910701] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.924072] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.924088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:4648 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.924094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.936841] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.936858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:486 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.936864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.950690] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.950706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:8126 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.950712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:62 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.960537] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.960553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:4523 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.960560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.973308] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.973324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:19795 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.973331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:57 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:02.987154] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:02.987171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:3468 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.906 [2024-11-06 09:08:02.987178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.906 [2024-11-06 09:08:03.000419] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.906 [2024-11-06 09:08:03.000436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:3329 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.907 [2024-11-06 09:08:03.000442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:12.907 [2024-11-06 09:08:03.013161] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:12.907 [2024-11-06 09:08:03.013177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:22990 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:12.907 [2024-11-06 09:08:03.013184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.167 [2024-11-06 09:08:03.025881] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.167 [2024-11-06 09:08:03.025899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:3332 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.167 [2024-11-06 09:08:03.025905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.167 [2024-11-06 09:08:03.036723] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.167 [2024-11-06 09:08:03.036739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:23510 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.167 [2024-11-06 09:08:03.036748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.050390] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.050406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:17269 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.050412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.061690] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.061706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:14168 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.061713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.075575] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.075591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:12527 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.075597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.088634] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.088650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:14982 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.088657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.101297] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.101314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:14920 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.101324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.113531] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.113547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:2370 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.113554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.125353] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.125369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:23867 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.125375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.135971] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.135987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:10060 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.135994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.150558] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.150575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:9241 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.150581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.162397] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.162414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:19829 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.162420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.174155] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.174171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:14459 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.174177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.186376] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.186392] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:459 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.186399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:112 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.200303] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.200320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:17651 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.200326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.213732] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.213755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:4811 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.213761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.227318] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.227335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:14775 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.227341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.239234] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.239251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:13715 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.239259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.249907] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.249924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4181 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.249930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.263337] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.263354] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:6926 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.263361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.168 [2024-11-06 09:08:03.277772] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.168 [2024-11-06 09:08:03.277788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23138 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.168 [2024-11-06 09:08:03.277794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.289205] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.289221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:6183 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.289227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.301741] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.301761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:24364 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.301767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.313812] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.313829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:21752 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.313835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.327514] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.327531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:9317 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.327537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.339636] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.339652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:1785 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.339658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.352027] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.352044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:13783 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.352050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.363457] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.363473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:7729 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.363480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.376760] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.376777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:17426 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.376783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.389310] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.389327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:14695 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.389333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.401412] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.401428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:9760 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.401435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.414050] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.414066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:15699 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.414073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.426634] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.426650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:14897 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.426660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.437955] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.437971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:15047 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.437977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.450910] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.450926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:1352 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.450933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.463625] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.463643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:21480 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.463649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.476047] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.476064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:3574 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.476071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.489832] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.489849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:23639 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.489856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.501218] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.501235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:508 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.501241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.512394] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.512410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:24680 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.512417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:48 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.526669] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.526686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:23303 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.526693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.430 [2024-11-06 09:08:03.539586] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xf5b200) 00:32:13.430 [2024-11-06 09:08:03.539606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11083 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:13.430 [2024-11-06 09:08:03.539613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:13.691 20170.50 IOPS, 78.79 MiB/s 00:32:13.691 Latency(us) 00:32:13.691 [2024-11-06T08:08:03.805Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:13.692 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:32:13.692 nvme0n1 : 2.00 20200.82 78.91 0.00 0.00 6330.82 2443.95 22391.47 00:32:13.692 [2024-11-06T08:08:03.805Z] =================================================================================================================== 00:32:13.692 [2024-11-06T08:08:03.805Z] Total : 20200.82 78.91 0.00 0.00 6330.82 2443.95 22391.47 00:32:13.692 { 00:32:13.692 "results": [ 00:32:13.692 { 00:32:13.692 "job": "nvme0n1", 00:32:13.692 "core_mask": "0x2", 00:32:13.692 "workload": "randread", 00:32:13.692 "status": "finished", 00:32:13.692 "queue_depth": 128, 00:32:13.692 "io_size": 4096, 00:32:13.692 "runtime": 2.003335, 00:32:13.692 "iops": 20200.815140752795, 00:32:13.692 "mibps": 78.9094341435656, 00:32:13.692 "io_failed": 0, 00:32:13.692 "io_timeout": 0, 00:32:13.692 "avg_latency_us": 6330.816088034463, 00:32:13.692 "min_latency_us": 2443.9466666666667, 00:32:13.692 "max_latency_us": 22391.466666666667 00:32:13.692 } 00:32:13.692 ], 00:32:13.692 "core_count": 1 00:32:13.692 } 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:32:13.692 | .driver_specific 00:32:13.692 | .nvme_error 00:32:13.692 | .status_code 00:32:13.692 | .command_transient_transport_error' 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 158 > 0 )) 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 2065151 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@950 -- # '[' -z 2065151 ']' 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # kill -0 2065151 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # uname 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2065151 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2065151' 00:32:13.692 killing process with pid 2065151 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@969 -- # kill 2065151 00:32:13.692 Received shutdown signal, test time was about 2.000000 seconds 00:32:13.692 00:32:13.692 Latency(us) 00:32:13.692 [2024-11-06T08:08:03.805Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:13.692 [2024-11-06T08:08:03.805Z] =================================================================================================================== 00:32:13.692 [2024-11-06T08:08:03.805Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:13.692 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@974 -- # wait 2065151 00:32:13.952 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@109 -- # run_bperf_err randread 131072 16 00:32:13.952 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:32:13.952 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randread 00:32:13.952 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=131072 00:32:13.952 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=16 00:32:13.952 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=2065996 00:32:13.953 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 2065996 /var/tmp/bperf.sock 00:32:13.953 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@831 -- # '[' -z 2065996 ']' 00:32:13.953 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:32:13.953 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:13.953 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:32:13.953 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:32:13.953 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:13.953 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:13.953 09:08:03 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z 00:32:13.953 [2024-11-06 09:08:03.944121] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:13.953 [2024-11-06 09:08:03.944178] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2065996 ] 00:32:13.953 I/O size of 131072 is greater than zero copy threshold (65536). 00:32:13.953 Zero copy mechanism will not be used. 00:32:13.953 [2024-11-06 09:08:04.026085] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:13.953 [2024-11-06 09:08:04.055025] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:14.893 09:08:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:14.893 09:08:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # return 0 00:32:14.893 09:08:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:32:14.893 09:08:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:32:14.893 09:08:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:32:14.893 09:08:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:14.893 09:08:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:14.893 09:08:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:14.893 09:08:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:14.893 09:08:04 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:15.154 nvme0n1 00:32:15.154 09:08:05 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:32:15.154 09:08:05 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:15.154 09:08:05 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:15.154 09:08:05 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:15.154 09:08:05 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:32:15.154 09:08:05 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:32:15.414 I/O size of 131072 is greater than zero copy threshold (65536). 00:32:15.414 Zero copy mechanism will not be used. 00:32:15.414 Running I/O for 2 seconds... 00:32:15.414 [2024-11-06 09:08:05.310568] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.414 [2024-11-06 09:08:05.310605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.414 [2024-11-06 09:08:05.310614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.414 [2024-11-06 09:08:05.321955] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.414 [2024-11-06 09:08:05.321977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.414 [2024-11-06 09:08:05.321985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.414 [2024-11-06 09:08:05.330413] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.330433] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.330440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.338775] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.338796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.338803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.350782] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.350801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.350807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.357941] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.357959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.357967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.368663] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.368682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.368694] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.378307] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.378326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.378333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.389051] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.389070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.389076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.397697] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.397716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.397723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.409139] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.409158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.409164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.422020] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.422038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.422046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.435235] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.435253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.435260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.448253] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.448271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.448278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.460046] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.460064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.460071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.471184] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.471209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.471216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.481740] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.481762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.481769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.487627] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.487645] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:64 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.487652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.492693] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.492711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.492717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.499372] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.499390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.499397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.506094] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.506113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:20032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.506119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.415 [2024-11-06 09:08:05.516690] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.415 [2024-11-06 09:08:05.516708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:18560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.415 [2024-11-06 09:08:05.516715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.526444] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.526464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.526470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.537095] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.537112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:13760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.537118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.546923] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.546942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.546948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.557407] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.557425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.557432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.564420] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.564439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.564445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.573636] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.573655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.573661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.579198] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.579217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.579223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.589347] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.589366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.589373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.600282] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.600301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:8384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.600307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.611354] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.611372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:13824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.611379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.619662] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.619681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:9984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.619691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.630869] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.630888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.630894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.642463] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.642482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:21440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.642488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.653328] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.653346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:1472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.653353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.663400] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.663419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:19488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.663425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.671549] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.671568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.671575] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.679916] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.679934] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:6848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.679941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.690191] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.690210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:3744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.690217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.701263] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.701282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:19872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.701289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.707549] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.707572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.707578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.715906] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.677 [2024-11-06 09:08:05.715924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:9184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.677 [2024-11-06 09:08:05.715930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.677 [2024-11-06 09:08:05.725866] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.678 [2024-11-06 09:08:05.725884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.678 [2024-11-06 09:08:05.725890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.678 [2024-11-06 09:08:05.735060] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.678 [2024-11-06 09:08:05.735079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:14432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.678 [2024-11-06 09:08:05.735086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.678 [2024-11-06 09:08:05.745535] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.678 [2024-11-06 09:08:05.745554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:15424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.678 [2024-11-06 09:08:05.745560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.678 [2024-11-06 09:08:05.756531] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.678 [2024-11-06 09:08:05.756550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:13088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.678 [2024-11-06 09:08:05.756556] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.678 [2024-11-06 09:08:05.768055] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.678 [2024-11-06 09:08:05.768074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:20288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.678 [2024-11-06 09:08:05.768081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.678 [2024-11-06 09:08:05.779657] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.678 [2024-11-06 09:08:05.779676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:3936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.678 [2024-11-06 09:08:05.779683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.788610] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.788630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:18112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.788640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.796858] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.796877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.796883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.806609] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.806628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:12256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.806634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.815795] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.815813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:18304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.815820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.826957] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.826976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:21856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.826982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.836790] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.836809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.836816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.845925] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.845943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:10048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.845950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.856630] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.856649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:20416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.856655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.868126] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.868145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:5888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.868152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.877024] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.877046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:4608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.877053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.889105] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.889124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:3808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.889130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.901820] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.901838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:13632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.939 [2024-11-06 09:08:05.901845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.939 [2024-11-06 09:08:05.913803] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.939 [2024-11-06 09:08:05.913822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:05.913829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:05.925661] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:05.925681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:2176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:05.925688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:05.937895] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:05.937914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:05.937921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:05.949916] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:05.949935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:18016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:05.949942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:05.961944] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:05.961964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:20992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:05.961970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:05.975332] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:05.975351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:05.975358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:05.986558] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:05.986577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:21152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:05.986583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:05.998524] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:05.998542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:1152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:05.998548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:06.011187] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:06.011206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:13920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:06.011213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:06.023921] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:06.023940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:20608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:06.023947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:06.036694] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:06.036712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:06.036719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:15.940 [2024-11-06 09:08:06.049721] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:15.940 [2024-11-06 09:08:06.049740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:14016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:15.940 [2024-11-06 09:08:06.049750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.061236] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.061256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.061262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.073317] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.073337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:15360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.073343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.084764] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.084782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:19104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.084792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.097338] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.097356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.097363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.109878] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.109897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:13632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.109904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.122460] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.122478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:8448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.122484] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.133691] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.133710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.133716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.145442] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.145461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:2048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.145468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.157765] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.157784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:12960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.157790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.167408] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.167427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.167434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.177443] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.177461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:7776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.177468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.187961] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.187983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:18240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.187990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.197583] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.197601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:21440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.197608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.207648] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.207668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.207674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.217228] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.217246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:24576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.217253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.228233] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.228251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:3616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.228258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.238567] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.238586] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:12064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.238592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.248635] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.248653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:6720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.248660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.260298] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.260316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:4448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.260323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.270348] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.270367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:2400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.270373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.280854] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.280872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:21824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.280878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.201 [2024-11-06 09:08:06.290402] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.290420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:3232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.290427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.201 2957.00 IOPS, 369.62 MiB/s [2024-11-06T08:08:06.314Z] [2024-11-06 09:08:06.302659] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.201 [2024-11-06 09:08:06.302678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:24032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.201 [2024-11-06 09:08:06.302684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.462 [2024-11-06 09:08:06.313792] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.462 [2024-11-06 09:08:06.313810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:14400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.462 [2024-11-06 09:08:06.313817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.462 [2024-11-06 09:08:06.322618] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.462 [2024-11-06 09:08:06.322637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:11392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.462 [2024-11-06 09:08:06.322643] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.462 [2024-11-06 09:08:06.331764] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.462 [2024-11-06 09:08:06.331783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:11808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.462 [2024-11-06 09:08:06.331789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.462 [2024-11-06 09:08:06.341643] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.462 [2024-11-06 09:08:06.341661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.341667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.352350] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.352369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:22400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.352375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.361897] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.361916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:12640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.361926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.370932] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.370950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:2752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.370957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.381325] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.381344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:11296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.381350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.391347] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.391366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:2592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.391373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.398246] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.398265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:22080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.398271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.406061] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.406080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:4256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.406088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.413780] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.413799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:22208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.413805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.420670] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.420689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:18688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.420695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.428689] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.428708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:13184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.428714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.434114] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.434131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.434138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.441226] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.441245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:5728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.441251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.449418] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.449437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:22272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.449443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.456476] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.456495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:11200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.456501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.461921] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.461939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:5376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.461946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.471878] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.471896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:18528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.471902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.480597] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.480616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:16736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.480622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.489430] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.489449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:22336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.489455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.497528] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.497547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:5536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.497558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.505554] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.505573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:7200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.505579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.511940] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.511959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:19104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.511967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.520963] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.520981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:4256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.520988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.523895] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.523913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:9952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.523919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.533179] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.533197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:3872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.533204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.541995] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.542013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:1056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.542020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.463 [2024-11-06 09:08:06.551322] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.463 [2024-11-06 09:08:06.551341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:15424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.463 [2024-11-06 09:08:06.551347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.464 [2024-11-06 09:08:06.558585] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.464 [2024-11-06 09:08:06.558604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:13600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.464 [2024-11-06 09:08:06.558610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.464 [2024-11-06 09:08:06.566879] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.464 [2024-11-06 09:08:06.566900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:20672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.464 [2024-11-06 09:08:06.566906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.724 [2024-11-06 09:08:06.578310] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.724 [2024-11-06 09:08:06.578329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.724 [2024-11-06 09:08:06.578335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.724 [2024-11-06 09:08:06.584215] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.724 [2024-11-06 09:08:06.584233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:24672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.724 [2024-11-06 09:08:06.584240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.724 [2024-11-06 09:08:06.591807] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.724 [2024-11-06 09:08:06.591825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:14816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.591831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.603435] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.603454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:22176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.603461] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.615772] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.615791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:7360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.615797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.627731] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.627753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.627760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.640407] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.640425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:2400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.640432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.651956] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.651975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:21504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.651981] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.662135] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.662153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:9184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.662159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.672783] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.672801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:25184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.672808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.682883] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.682901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:7520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.682908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.691800] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.691818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:5696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.691824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.701598] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.701616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:15072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.701623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.709940] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.709959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.709966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.717239] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.717258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.717264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.722921] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.722939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:15616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.722946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.729104] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.729123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:13440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.729135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.739994] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.740013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:13920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.740019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.751937] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.751956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.751962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.758090] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.758108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:5568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.758114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.767117] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.767135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:8640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.767142] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.772743] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.772766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.772772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.780840] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.780859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:10848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.780865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.789244] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.789263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.789269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.795251] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.795269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:12736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.795276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.801086] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.801107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:25504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.801114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.813346] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.813365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:12288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.813371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.820759] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.820779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.820786] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.725 [2024-11-06 09:08:06.829568] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.725 [2024-11-06 09:08:06.829587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:2336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.725 [2024-11-06 09:08:06.829593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.986 [2024-11-06 09:08:06.839964] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.986 [2024-11-06 09:08:06.839982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.986 [2024-11-06 09:08:06.839988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.986 [2024-11-06 09:08:06.852461] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.852479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:19968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.852486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.865451] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.865470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:6944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.865477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.878523] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.878542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:16128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.878549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.891440] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.891459] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:3520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.891465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.904666] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.904685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:4608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.904691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.917614] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.917633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:6560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.917639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.929882] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.929901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:20576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.929907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.943090] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.943108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:1856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.943115] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.955321] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.955339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.955346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.967113] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.967132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:10016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.967138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.976162] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.976181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.976188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.985284] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.985303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:24608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.985310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:06.995571] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:06.995593] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:8864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:06.995600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.004660] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.004679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:18752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.004685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.009721] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.009741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:24320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.009752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.018259] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.018278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.018284] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.024145] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.024165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:14592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.024172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.029957] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.029976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:2848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.029982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.039379] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.039398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:24640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.039405] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.047324] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.047343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:14912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.047350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.052879] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.052898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:14688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.052905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.061565] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.061583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:15968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.061590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.070025] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.070043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:9344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.070050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.076142] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.076161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:5824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.076168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.084904] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.084922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:13152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.084928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:16.987 [2024-11-06 09:08:07.093821] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:16.987 [2024-11-06 09:08:07.093840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:5760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:16.987 [2024-11-06 09:08:07.093847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:17.249 [2024-11-06 09:08:07.104614] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.249 [2024-11-06 09:08:07.104634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:17152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.249 [2024-11-06 09:08:07.104640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:17.249 [2024-11-06 09:08:07.114738] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.249 [2024-11-06 09:08:07.114761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.249 [2024-11-06 09:08:07.114768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:17.249 [2024-11-06 09:08:07.121454] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.249 [2024-11-06 09:08:07.121473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:19808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.249 [2024-11-06 09:08:07.121479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:17.249 [2024-11-06 09:08:07.130578] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.249 [2024-11-06 09:08:07.130597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:15584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.249 [2024-11-06 09:08:07.130607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:17.249 [2024-11-06 09:08:07.139687] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.249 [2024-11-06 09:08:07.139706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:1152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.249 [2024-11-06 09:08:07.139712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:17.249 [2024-11-06 09:08:07.145881] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.249 [2024-11-06 09:08:07.145899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.249 [2024-11-06 09:08:07.145905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:17.249 [2024-11-06 09:08:07.155367] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.249 [2024-11-06 09:08:07.155385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:11072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.249 [2024-11-06 09:08:07.155392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:17.249 [2024-11-06 09:08:07.163735] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.249 [2024-11-06 09:08:07.163759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:11488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.249 [2024-11-06 09:08:07.163766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:17.249 [2024-11-06 09:08:07.172388] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.249 [2024-11-06 09:08:07.172407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:16384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.249 [2024-11-06 09:08:07.172414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.182375] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.182394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:14400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.182400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.191374] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.191393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:12960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.191400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.198076] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.198095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:5152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.198101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.206447] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.206468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:13920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.206475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.213756] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.213774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:1632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.213781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.222091] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.222110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:12192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.222116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.231042] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.231060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.231067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.239751] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.239769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:12960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.239775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.243310] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.243329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:10016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.243337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.251253] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.251271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:6208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.251278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.259076] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.259095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:21344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.259101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.269980] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.269999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:24960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.270005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:17.250 [2024-11-06 09:08:07.281900] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.281919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:8832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.281925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:17.250 3203.00 IOPS, 400.38 MiB/s [2024-11-06T08:08:07.363Z] [2024-11-06 09:08:07.295913] nvme_tcp.c:1365:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e61620) 00:32:17.250 [2024-11-06 09:08:07.295929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:2176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:17.250 [2024-11-06 09:08:07.295936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:17.250 00:32:17.250 Latency(us) 00:32:17.250 [2024-11-06T08:08:07.363Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:17.250 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:32:17.250 nvme0n1 : 2.00 3206.80 400.85 0.00 0.00 4984.65 675.84 15947.09 00:32:17.250 [2024-11-06T08:08:07.363Z] =================================================================================================================== 00:32:17.250 [2024-11-06T08:08:07.363Z] Total : 3206.80 400.85 0.00 0.00 4984.65 675.84 15947.09 00:32:17.250 { 00:32:17.250 "results": [ 00:32:17.250 { 00:32:17.250 "job": "nvme0n1", 00:32:17.250 "core_mask": "0x2", 00:32:17.250 "workload": "randread", 00:32:17.250 "status": "finished", 00:32:17.250 "queue_depth": 16, 00:32:17.250 "io_size": 131072, 00:32:17.250 "runtime": 2.002618, 00:32:17.250 "iops": 3206.8022957948046, 00:32:17.250 "mibps": 400.8502869743506, 00:32:17.250 "io_failed": 0, 00:32:17.250 "io_timeout": 0, 00:32:17.250 "avg_latency_us": 4984.65352019101, 00:32:17.250 "min_latency_us": 675.84, 00:32:17.250 "max_latency_us": 15947.093333333334 00:32:17.250 } 00:32:17.250 ], 00:32:17.250 "core_count": 1 00:32:17.250 } 00:32:17.250 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:32:17.250 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:32:17.250 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:32:17.250 | .driver_specific 00:32:17.250 | .nvme_error 00:32:17.250 | .status_code 00:32:17.250 | .command_transient_transport_error' 00:32:17.250 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 207 > 0 )) 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 2065996 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@950 -- # '[' -z 2065996 ']' 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # kill -0 2065996 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # uname 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2065996 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2065996' 00:32:17.511 killing process with pid 2065996 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@969 -- # kill 2065996 00:32:17.511 Received shutdown signal, test time was about 2.000000 seconds 00:32:17.511 00:32:17.511 Latency(us) 00:32:17.511 [2024-11-06T08:08:07.624Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:17.511 [2024-11-06T08:08:07.624Z] =================================================================================================================== 00:32:17.511 [2024-11-06T08:08:07.624Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:17.511 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@974 -- # wait 2065996 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@114 -- # run_bperf_err randwrite 4096 128 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randwrite 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=4096 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=128 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=2066801 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 2066801 /var/tmp/bperf.sock 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@831 -- # '[' -z 2066801 ']' 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:32:17.772 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:17.772 09:08:07 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:17.772 [2024-11-06 09:08:07.713337] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:17.772 [2024-11-06 09:08:07.713396] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2066801 ] 00:32:17.772 [2024-11-06 09:08:07.797105] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:17.772 [2024-11-06 09:08:07.826963] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:18.712 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:18.712 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # return 0 00:32:18.712 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:32:18.712 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:32:18.712 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:32:18.712 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:18.712 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:18.712 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:18.712 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:18.712 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:18.973 nvme0n1 00:32:18.973 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:32:18.973 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:18.973 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:18.973 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:18.973 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:32:18.973 09:08:08 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:32:18.973 Running I/O for 2 seconds... 00:32:18.973 [2024-11-06 09:08:09.078782] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eeb760 00:32:18.973 [2024-11-06 09:08:09.080427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:33 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:18.973 [2024-11-06 09:08:09.080455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.088475] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2d80 00:32:19.234 [2024-11-06 09:08:09.089439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:5646 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.234 [2024-11-06 09:08:09.089456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.101606] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee5220 00:32:19.234 [2024-11-06 09:08:09.102727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:25093 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.234 [2024-11-06 09:08:09.102743] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.113705] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efa7d8 00:32:19.234 [2024-11-06 09:08:09.114805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:1430 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.234 [2024-11-06 09:08:09.114822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.125715] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efa7d8 00:32:19.234 [2024-11-06 09:08:09.126816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:3697 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.234 [2024-11-06 09:08:09.126832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:100 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.137700] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efa7d8 00:32:19.234 [2024-11-06 09:08:09.138806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:4243 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.234 [2024-11-06 09:08:09.138822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.149682] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efa7d8 00:32:19.234 [2024-11-06 09:08:09.150775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:3853 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.234 [2024-11-06 09:08:09.150791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.161658] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efa7d8 00:32:19.234 [2024-11-06 09:08:09.162757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:24705 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.234 [2024-11-06 09:08:09.162773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.173564] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee95a0 00:32:19.234 [2024-11-06 09:08:09.174659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:3855 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.234 [2024-11-06 09:08:09.174674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.185558] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee95a0 00:32:19.234 [2024-11-06 09:08:09.186606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:19104 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.234 [2024-11-06 09:08:09.186621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.196912] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee5658 00:32:19.234 [2024-11-06 09:08:09.197946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:16238 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.234 [2024-11-06 09:08:09.197962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:32:19.234 [2024-11-06 09:08:09.210055] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef0350 00:32:19.235 [2024-11-06 09:08:09.211312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:16809 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.211328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.222009] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2d80 00:32:19.235 [2024-11-06 09:08:09.223246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:5649 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.223262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.233967] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2d80 00:32:19.235 [2024-11-06 09:08:09.235234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:11898 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.235250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.245935] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2d80 00:32:19.235 [2024-11-06 09:08:09.247189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:8188 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.247208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.257883] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2d80 00:32:19.235 [2024-11-06 09:08:09.259122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:12235 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.259138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.269850] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2d80 00:32:19.235 [2024-11-06 09:08:09.271105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:17026 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.271121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.281800] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2d80 00:32:19.235 [2024-11-06 09:08:09.283005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:16599 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.283020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:111 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.292966] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef0bc0 00:32:19.235 [2024-11-06 09:08:09.294189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:18295 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.294204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:112 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.307167] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2510 00:32:19.235 [2024-11-06 09:08:09.309023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:13075 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.309038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.317966] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef81e0 00:32:19.235 [2024-11-06 09:08:09.319323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:12379 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.319338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.330125] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef3a28 00:32:19.235 [2024-11-06 09:08:09.331489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:23259 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.331505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:32:19.235 [2024-11-06 09:08:09.341252] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef4298 00:32:19.235 [2024-11-06 09:08:09.342599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:2193 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.235 [2024-11-06 09:08:09.342614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.353143] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef4b08 00:32:19.496 [2024-11-06 09:08:09.354440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:7969 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.354456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.365055] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef9b30 00:32:19.496 [2024-11-06 09:08:09.366376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:4742 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.366392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.377755] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef9b30 00:32:19.496 [2024-11-06 09:08:09.379033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:4125 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.379049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.389681] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef4b08 00:32:19.496 [2024-11-06 09:08:09.390970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:15338 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.390985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.401719] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eed920 00:32:19.496 [2024-11-06 09:08:09.403052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:3328 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.403068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.412928] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee88f8 00:32:19.496 [2024-11-06 09:08:09.414230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:19676 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.414246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.425633] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee88f8 00:32:19.496 [2024-11-06 09:08:09.426949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:7330 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.426966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.437583] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee88f8 00:32:19.496 [2024-11-06 09:08:09.438915] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:13846 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.438931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.449556] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee88f8 00:32:19.496 [2024-11-06 09:08:09.450871] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:7472 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.450887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.460764] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec408 00:32:19.496 [2024-11-06 09:08:09.462062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:10790 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.462077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.473471] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec408 00:32:19.496 [2024-11-06 09:08:09.474779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:8927 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.474795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.485445] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec408 00:32:19.496 [2024-11-06 09:08:09.486757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:6550 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.486772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.497395] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec408 00:32:19.496 [2024-11-06 09:08:09.498701] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:18844 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.498718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.509353] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec408 00:32:19.496 [2024-11-06 09:08:09.510668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:12830 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.510683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.521290] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec408 00:32:19.496 [2024-11-06 09:08:09.522599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:22970 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.522614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.533251] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec408 00:32:19.496 [2024-11-06 09:08:09.534556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:1242 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.534572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.545196] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec408 00:32:19.496 [2024-11-06 09:08:09.546499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:7076 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.546515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.558689] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec408 00:32:19.496 [2024-11-06 09:08:09.560646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:2927 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.496 [2024-11-06 09:08:09.560664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:19.496 [2024-11-06 09:08:09.570777] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eeea00 00:32:19.496 [2024-11-06 09:08:09.572678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:5770 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.497 [2024-11-06 09:08:09.572693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:32:19.497 [2024-11-06 09:08:09.580784] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efc128 00:32:19.497 [2024-11-06 09:08:09.582219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:20549 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.497 [2024-11-06 09:08:09.582234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:32:19.497 [2024-11-06 09:08:09.592705] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efb8b8 00:32:19.497 [2024-11-06 09:08:09.594104] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:13843 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.497 [2024-11-06 09:08:09.594119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:32:19.497 [2024-11-06 09:08:09.605436] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee73e0 00:32:19.497 [2024-11-06 09:08:09.606822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:13744 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.497 [2024-11-06 09:08:09.606837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.618984] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee6300 00:32:19.757 [2024-11-06 09:08:09.621056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:20241 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.621073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.628574] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef92c0 00:32:19.757 [2024-11-06 09:08:09.629976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:20518 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.629991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:100 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.641255] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef92c0 00:32:19.757 [2024-11-06 09:08:09.642659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:19241 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.642675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.653214] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef92c0 00:32:19.757 [2024-11-06 09:08:09.654636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:17260 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.654651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.665164] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef92c0 00:32:19.757 [2024-11-06 09:08:09.666552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:22408 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.666568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:89 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.677116] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef92c0 00:32:19.757 [2024-11-06 09:08:09.678535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:12994 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.678551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.689062] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef92c0 00:32:19.757 [2024-11-06 09:08:09.690464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:10897 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.690481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.701009] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef92c0 00:32:19.757 [2024-11-06 09:08:09.702407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:10267 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.702423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.712967] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef92c0 00:32:19.757 [2024-11-06 09:08:09.714372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:13458 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.714388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.724919] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef92c0 00:32:19.757 [2024-11-06 09:08:09.726327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:19017 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.726343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.736299] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2510 00:32:19.757 [2024-11-06 09:08:09.737688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:11115 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.737704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.748192] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2d80 00:32:19.757 [2024-11-06 09:08:09.749567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:14889 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.749582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.760102] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef81e0 00:32:19.757 [2024-11-06 09:08:09.761459] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:1349 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.761474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.771985] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef35f0 00:32:19.757 [2024-11-06 09:08:09.773331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:6875 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.773346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.784717] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef35f0 00:32:19.757 [2024-11-06 09:08:09.786077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:1602 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.786092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.798209] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef35f0 00:32:19.757 [2024-11-06 09:08:09.800182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:12526 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.800197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.808670] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee8d30 00:32:19.757 [2024-11-06 09:08:09.809995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:4392 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.810010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:114 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.819849] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef4298 00:32:19.757 [2024-11-06 09:08:09.821185] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:15500 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.821200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.832531] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee95a0 00:32:19.757 [2024-11-06 09:08:09.833863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:17174 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.757 [2024-11-06 09:08:09.833879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:32:19.757 [2024-11-06 09:08:09.843641] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee9e10 00:32:19.758 [2024-11-06 09:08:09.844956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:10931 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.758 [2024-11-06 09:08:09.844971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:32:19.758 [2024-11-06 09:08:09.856333] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee9e10 00:32:19.758 [2024-11-06 09:08:09.857650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:22064 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:19.758 [2024-11-06 09:08:09.857665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:32:19.758 [2024-11-06 09:08:09.868321] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee9e10 00:32:20.018 [2024-11-06 09:08:09.869633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:16733 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.018 [2024-11-06 09:08:09.869652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:54 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:32:20.018 [2024-11-06 09:08:09.880271] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee9e10 00:32:20.018 [2024-11-06 09:08:09.881584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:1875 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:09.881599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:09.892174] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eee5c8 00:32:20.019 [2024-11-06 09:08:09.893482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:16050 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:09.893498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:09.904131] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eee5c8 00:32:20.019 [2024-11-06 09:08:09.905432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:4801 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:09.905448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:09.916098] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eee5c8 00:32:20.019 [2024-11-06 09:08:09.917376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:18238 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:09.917392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:09.928050] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eee5c8 00:32:20.019 [2024-11-06 09:08:09.929350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:20438 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:09.929366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:09.940003] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eee5c8 00:32:20.019 [2024-11-06 09:08:09.941263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:10005 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:09.941278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:09.951904] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef7100 00:32:20.019 [2024-11-06 09:08:09.953156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:20401 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:09.953171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:09.963821] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eee5c8 00:32:20.019 [2024-11-06 09:08:09.965077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:4625 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:09.965092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:09.975810] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eeaab8 00:32:20.019 [2024-11-06 09:08:09.977102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:2797 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:09.977121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:09.987762] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016edfdc0 00:32:20.019 [2024-11-06 09:08:09.989042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:3829 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:09.989058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:09.999750] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef7100 00:32:20.019 [2024-11-06 09:08:10.000951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:6727 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.000967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:10.013694] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016edfdc0 00:32:20.019 [2024-11-06 09:08:10.015629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:23976 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.015645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:54 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:10.024104] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef6020 00:32:20.019 [2024-11-06 09:08:10.025370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:8791 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.025386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:40 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:10.036072] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef6020 00:32:20.019 [2024-11-06 09:08:10.037202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:8284 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.037218] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:10.048127] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eef270 00:32:20.019 [2024-11-06 09:08:10.049366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:14904 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.049382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:10.060091] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee3060 00:32:20.019 [2024-11-06 09:08:10.061351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:11205 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.061367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:32:20.019 21267.00 IOPS, 83.07 MiB/s [2024-11-06T08:08:10.132Z] [2024-11-06 09:08:10.073586] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee3060 00:32:20.019 [2024-11-06 09:08:10.075495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:23871 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.075510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:10.083242] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee5a90 00:32:20.019 [2024-11-06 09:08:10.084449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:20035 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.084465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:70 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:10.095152] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee5220 00:32:20.019 [2024-11-06 09:08:10.096367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:13223 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.096382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:10.107161] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eed4e8 00:32:20.019 [2024-11-06 09:08:10.108343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:10776 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.108358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:32:20.019 [2024-11-06 09:08:10.120029] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eddc00 00:32:20.019 [2024-11-06 09:08:10.121243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:24132 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.019 [2024-11-06 09:08:10.121259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.132009] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee49b0 00:32:20.280 [2024-11-06 09:08:10.133207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:21140 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.133222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.143973] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef0788 00:32:20.280 [2024-11-06 09:08:10.145174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:17710 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.145190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.155972] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee5220 00:32:20.280 [2024-11-06 09:08:10.157155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:12356 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.157171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.167953] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef6020 00:32:20.280 [2024-11-06 09:08:10.169096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:14707 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.169113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.179856] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef0788 00:32:20.280 [2024-11-06 09:08:10.181000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:13351 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.181016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.191001] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee5a90 00:32:20.280 [2024-11-06 09:08:10.192121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:20942 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.192136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:112 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.203705] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eeff18 00:32:20.280 [2024-11-06 09:08:10.204861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:15833 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.204876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.217224] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eeee38 00:32:20.280 [2024-11-06 09:08:10.219007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:22091 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.219022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.229074] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eef6a8 00:32:20.280 [2024-11-06 09:08:10.230828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:4295 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.230844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.239448] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eeff18 00:32:20.280 [2024-11-06 09:08:10.240548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:11007 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.240563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.252845] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef7538 00:32:20.280 [2024-11-06 09:08:10.254602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:5210 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.254618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.263241] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef0ff8 00:32:20.280 [2024-11-06 09:08:10.264347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6583 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.264363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.275213] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef0ff8 00:32:20.280 [2024-11-06 09:08:10.276282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:20269 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.276298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.288682] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef0ff8 00:32:20.280 [2024-11-06 09:08:10.290398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:22182 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.290416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.300539] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eddc00 00:32:20.280 [2024-11-06 09:08:10.302243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:23865 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.302258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.310190] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2d80 00:32:20.280 [2024-11-06 09:08:10.311255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:9238 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.311270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.322053] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2510 00:32:20.280 [2024-11-06 09:08:10.323096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:7427 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.323112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.333921] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef1ca0 00:32:20.280 [2024-11-06 09:08:10.334922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:13592 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.334938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.345801] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee23b8 00:32:20.280 [2024-11-06 09:08:10.346765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:13268 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.346780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.360010] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efa7d8 00:32:20.280 [2024-11-06 09:08:10.361656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:7990 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.361671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.369599] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eea248 00:32:20.280 [2024-11-06 09:08:10.370568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:13538 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.370583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:48 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:32:20.280 [2024-11-06 09:08:10.382299] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee2c28 00:32:20.280 [2024-11-06 09:08:10.383250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:8989 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.280 [2024-11-06 09:08:10.383265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:32:20.541 [2024-11-06 09:08:10.394256] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efac10 00:32:20.541 [2024-11-06 09:08:10.395198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:13184 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.541 [2024-11-06 09:08:10.395214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:32:20.541 [2024-11-06 09:08:10.407744] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016edf988 00:32:20.541 [2024-11-06 09:08:10.409364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:21587 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.541 [2024-11-06 09:08:10.409379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:32:20.541 [2024-11-06 09:08:10.417324] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eeaef0 00:32:20.541 [2024-11-06 09:08:10.418266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:22611 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.541 [2024-11-06 09:08:10.418281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:32:20.541 [2024-11-06 09:08:10.430134] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee1710 00:32:20.541 [2024-11-06 09:08:10.431039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:25574 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.541 [2024-11-06 09:08:10.431055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:40 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:32:20.541 [2024-11-06 09:08:10.442111] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef1868 00:32:20.541 [2024-11-06 09:08:10.443025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:21885 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.541 [2024-11-06 09:08:10.443040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:32:20.541 [2024-11-06 09:08:10.455581] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee88f8 00:32:20.541 [2024-11-06 09:08:10.457184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:17947 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.541 [2024-11-06 09:08:10.457199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:32:20.541 [2024-11-06 09:08:10.465979] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee8088 00:32:20.541 [2024-11-06 09:08:10.466883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:11729 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.466898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.477144] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef5378 00:32:20.542 [2024-11-06 09:08:10.478022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:8988 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.478038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:64 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.489891] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef4f40 00:32:20.542 [2024-11-06 09:08:10.490775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:18174 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.490791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.503411] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef3e60 00:32:20.542 [2024-11-06 09:08:10.504981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:13996 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.504996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.513868] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efb480 00:32:20.542 [2024-11-06 09:08:10.514808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:9924 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.514823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.525770] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef46d0 00:32:20.542 [2024-11-06 09:08:10.526682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:17108 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.526697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.537731] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef46d0 00:32:20.542 [2024-11-06 09:08:10.538657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:12067 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.538673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.549678] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef46d0 00:32:20.542 [2024-11-06 09:08:10.550597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:20884 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.550612] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.561638] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef46d0 00:32:20.542 [2024-11-06 09:08:10.562523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:6535 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.562541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.572820] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee7c50 00:32:20.542 [2024-11-06 09:08:10.573716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6546 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.573731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.587073] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee7c50 00:32:20.542 [2024-11-06 09:08:10.588633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:20733 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.588649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.598990] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eecc78 00:32:20.542 [2024-11-06 09:08:10.600499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:16847 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.600518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.610886] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee6738 00:32:20.542 [2024-11-06 09:08:10.612396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:20322 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.612412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.621278] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee6300 00:32:20.542 [2024-11-06 09:08:10.622131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:13389 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.622148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:120 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.633245] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee6300 00:32:20.542 [2024-11-06 09:08:10.634100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:5519 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.634116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:32:20.542 [2024-11-06 09:08:10.645207] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee6300 00:32:20.542 [2024-11-06 09:08:10.646021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:3019 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.542 [2024-11-06 09:08:10.646036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.658706] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee6300 00:32:20.803 [2024-11-06 09:08:10.660205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:24206 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.660221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.671395] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee6738 00:32:20.803 [2024-11-06 09:08:10.672849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:3178 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.672864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.683322] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee6738 00:32:20.803 [2024-11-06 09:08:10.684765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:16842 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.684780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.695332] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec840 00:32:20.803 [2024-11-06 09:08:10.696828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:870 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.696844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.707320] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eec840 00:32:20.803 [2024-11-06 09:08:10.708774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:1080 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.708791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.719275] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee6b70 00:32:20.803 [2024-11-06 09:08:10.720710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:18879 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.720726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:119 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.732965] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efb8b8 00:32:20.803 [2024-11-06 09:08:10.735069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:18575 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.735085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.744836] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eecc78 00:32:20.803 [2024-11-06 09:08:10.746918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:15854 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.746934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.755200] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee7c50 00:32:20.803 [2024-11-06 09:08:10.756618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:24182 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.756634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.768715] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efe2e8 00:32:20.803 [2024-11-06 09:08:10.770807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:21648 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.770823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.779146] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee27f0 00:32:20.803 [2024-11-06 09:08:10.780588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:5295 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.780604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.791106] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee27f0 00:32:20.803 [2024-11-06 09:08:10.792501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:14367 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.792517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.803095] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee27f0 00:32:20.803 [2024-11-06 09:08:10.804492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:18002 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.803 [2024-11-06 09:08:10.804509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:32:20.803 [2024-11-06 09:08:10.815052] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee27f0 00:32:20.804 [2024-11-06 09:08:10.816458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:12181 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.804 [2024-11-06 09:08:10.816473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:32:20.804 [2024-11-06 09:08:10.828490] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ede8a8 00:32:20.804 [2024-11-06 09:08:10.830568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:8942 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.804 [2024-11-06 09:08:10.830584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:32:20.804 [2024-11-06 09:08:10.838891] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efdeb0 00:32:20.804 [2024-11-06 09:08:10.840289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:16423 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.804 [2024-11-06 09:08:10.840304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:32:20.804 [2024-11-06 09:08:10.850030] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efb048 00:32:20.804 [2024-11-06 09:08:10.851385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:9670 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.804 [2024-11-06 09:08:10.851401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:32:20.804 [2024-11-06 09:08:10.861955] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efa7d8 00:32:20.804 [2024-11-06 09:08:10.863285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:1521 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.804 [2024-11-06 09:08:10.863300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:32:20.804 [2024-11-06 09:08:10.872773] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef6020 00:32:20.804 [2024-11-06 09:08:10.873605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23548 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.804 [2024-11-06 09:08:10.873620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:32:20.804 [2024-11-06 09:08:10.884912] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efda78 00:32:20.804 [2024-11-06 09:08:10.885740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:5028 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.804 [2024-11-06 09:08:10.885759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:32:20.804 [2024-11-06 09:08:10.898455] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016efc998 00:32:20.804 [2024-11-06 09:08:10.899977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:3517 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.804 [2024-11-06 09:08:10.899992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:32:20.804 [2024-11-06 09:08:10.908862] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ee5ec8 00:32:20.804 [2024-11-06 09:08:10.909735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:17721 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:20.804 [2024-11-06 09:08:10.909758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:10.922352] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eeb328 00:32:21.065 [2024-11-06 09:08:10.923829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:25209 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:10.923844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:10.934458] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eea248 00:32:21.065 [2024-11-06 09:08:10.935944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:5539 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:10.935960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:123 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:10.945081] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eeaef0 00:32:21.065 [2024-11-06 09:08:10.946137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:23566 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:10.946152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:10.958761] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2d80 00:32:21.065 [2024-11-06 09:08:10.960436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:4982 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:10.960452] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:10.969095] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2510 00:32:21.065 [2024-11-06 09:08:10.970070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:15609 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:10.970086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:10.981105] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2510 00:32:21.065 [2024-11-06 09:08:10.982067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:7329 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:10.982083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:123 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:10.993095] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2510 00:32:21.065 [2024-11-06 09:08:10.994059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:6531 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:10.994075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:89 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:11.004276] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef8a50 00:32:21.065 [2024-11-06 09:08:11.005222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:22812 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:11.005237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:74 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:11.016958] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef2510 00:32:21.065 [2024-11-06 09:08:11.017918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:12283 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:11.017933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:11.028932] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef35f0 00:32:21.065 [2024-11-06 09:08:11.029865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:21646 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:11.029880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:11.040931] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016eee190 00:32:21.065 [2024-11-06 09:08:11.041885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:2007 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:11.041901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:11.052098] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef81e0 00:32:21.065 [2024-11-06 09:08:11.053020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:2830 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:11.053035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:32:21.065 [2024-11-06 09:08:11.063987] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb050) with pdu=0x200016ef7970 00:32:21.065 [2024-11-06 09:08:11.064916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:18418 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:21.065 [2024-11-06 09:08:11.064931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:32:21.065 21297.00 IOPS, 83.19 MiB/s 00:32:21.065 Latency(us) 00:32:21.065 [2024-11-06T08:08:11.178Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:21.065 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:32:21.065 nvme0n1 : 2.00 21312.90 83.25 0.00 0.00 5997.88 2293.76 14199.47 00:32:21.065 [2024-11-06T08:08:11.178Z] =================================================================================================================== 00:32:21.065 [2024-11-06T08:08:11.178Z] Total : 21312.90 83.25 0.00 0.00 5997.88 2293.76 14199.47 00:32:21.065 { 00:32:21.065 "results": [ 00:32:21.065 { 00:32:21.065 "job": "nvme0n1", 00:32:21.065 "core_mask": "0x2", 00:32:21.065 "workload": "randwrite", 00:32:21.065 "status": "finished", 00:32:21.065 "queue_depth": 128, 00:32:21.065 "io_size": 4096, 00:32:21.065 "runtime": 2.004514, 00:32:21.065 "iops": 21312.89679194059, 00:32:21.065 "mibps": 83.25350309351793, 00:32:21.065 "io_failed": 0, 00:32:21.065 "io_timeout": 0, 00:32:21.065 "avg_latency_us": 5997.877023235492, 00:32:21.065 "min_latency_us": 2293.76, 00:32:21.065 "max_latency_us": 14199.466666666667 00:32:21.065 } 00:32:21.065 ], 00:32:21.065 "core_count": 1 00:32:21.065 } 00:32:21.065 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:32:21.065 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:32:21.065 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:32:21.065 | .driver_specific 00:32:21.065 | .nvme_error 00:32:21.065 | .status_code 00:32:21.065 | .command_transient_transport_error' 00:32:21.065 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 167 > 0 )) 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 2066801 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@950 -- # '[' -z 2066801 ']' 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # kill -0 2066801 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # uname 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2066801 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2066801' 00:32:21.325 killing process with pid 2066801 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@969 -- # kill 2066801 00:32:21.325 Received shutdown signal, test time was about 2.000000 seconds 00:32:21.325 00:32:21.325 Latency(us) 00:32:21.325 [2024-11-06T08:08:11.438Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:21.325 [2024-11-06T08:08:11.438Z] =================================================================================================================== 00:32:21.325 [2024-11-06T08:08:11.438Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:21.325 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@974 -- # wait 2066801 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@115 -- # run_bperf_err randwrite 131072 16 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randwrite 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=131072 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=16 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=2067510 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 2067510 /var/tmp/bperf.sock 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@831 -- # '[' -z 2067510 ']' 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:32:21.586 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:21.586 09:08:11 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:21.586 [2024-11-06 09:08:11.489892] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:21.586 [2024-11-06 09:08:11.489947] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2067510 ] 00:32:21.586 I/O size of 131072 is greater than zero copy threshold (65536). 00:32:21.586 Zero copy mechanism will not be used. 00:32:21.586 [2024-11-06 09:08:11.574894] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:21.586 [2024-11-06 09:08:11.603519] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:22.528 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:22.528 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@864 -- # return 0 00:32:22.528 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:32:22.528 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:32:22.528 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:32:22.528 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:22.528 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:22.528 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:22.528 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:22.528 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:32:22.789 nvme0n1 00:32:22.789 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:32:22.789 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:22.789 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:22.789 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:22.789 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:32:22.789 09:08:12 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:32:23.050 I/O size of 131072 is greater than zero copy threshold (65536). 00:32:23.050 Zero copy mechanism will not be used. 00:32:23.050 Running I/O for 2 seconds... 00:32:23.050 [2024-11-06 09:08:12.967799] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:12.968151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:12.968179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:12.972862] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:12.973073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:12.973093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:12.976981] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:12.977187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:12.977204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:12.981203] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:12.981406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:12.981423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:12.987561] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:12.987894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:12.987911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:12.994579] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:12.994913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:96 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:12.994930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:13.001284] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:13.001487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:13.001503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:13.007342] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:13.007555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:13.007572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:13.012035] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:13.012237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:13.012253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:13.019619] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:13.019952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:13.019969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:13.026393] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:13.026716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:13.026734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:13.033502] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:13.033704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.050 [2024-11-06 09:08:13.033724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.050 [2024-11-06 09:08:13.038287] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.050 [2024-11-06 09:08:13.038490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.038506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.046741] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.046952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.046969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.054640] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.054961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.054978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.061795] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.061996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.062013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.071117] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.071530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.071546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.078139] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.078343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.078359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.087692] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.087896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.087913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.094122] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.094334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.094350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.103872] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.104079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.104096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.111495] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.111699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.111717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.115809] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.116012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.116028] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.121624] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.121830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.121847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.126258] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.126462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.126479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.133412] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.133605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.133622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.137607] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.137813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.137830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.141631] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.141836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.141852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.145923] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.146124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.146140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.149990] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.150192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.150209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.154458] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.154660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.154675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.051 [2024-11-06 09:08:13.158558] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.051 [2024-11-06 09:08:13.158763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.051 [2024-11-06 09:08:13.158779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.162515] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.162716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.162733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.166585] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.166789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.166806] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.170711] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.170920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.170936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.175143] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.175343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.175360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.179339] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.179667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.179683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.183706] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.183911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.183931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.190190] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.190381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.190397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.196021] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.196343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.196360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.203424] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.203704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.203722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.213158] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.213355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.213371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.222679] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.222876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.222892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.231476] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.231666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.231683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.240952] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.241335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.241352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.248628] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.248824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.248841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.254377] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.254572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.313 [2024-11-06 09:08:13.254589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.313 [2024-11-06 09:08:13.259802] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.313 [2024-11-06 09:08:13.259993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.260010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.266507] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.266698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.266715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.274478] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.274740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.274761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.280272] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.280461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.280479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.287097] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.287475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.287492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.294433] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.294690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.294706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.302179] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.302370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.302386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.310578] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.310776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.310793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.318010] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.318200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.318216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.324793] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.324982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.324998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.329099] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.329289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.329304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.333844] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.334036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.334052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.337972] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.338130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.338146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.343669] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.343968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.343986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.350550] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.350785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.350801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.355176] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.355364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.355380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.362209] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.362485] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.362505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.370142] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.370359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.370375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.374708] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.374900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.374916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.379059] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.379248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.379264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.384547] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.384815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.384832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.390109] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.390298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.390314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.394935] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.395126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.395143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.401081] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.401270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.401286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.408006] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.408194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.408210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.314 [2024-11-06 09:08:13.418565] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.314 [2024-11-06 09:08:13.418820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.314 [2024-11-06 09:08:13.418836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.427765] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.428072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22688 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.428088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.438093] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.438386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.438403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.442576] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.442771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.442787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.446423] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.446613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.446630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.450564] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.450759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.450776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.454614] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.454809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.454825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.459168] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.459358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.459374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.464889] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.465206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.465226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.469124] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.469313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.469330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.476924] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.477120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.477136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.481107] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.481298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.481315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.484933] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.485123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.485140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.488744] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.488941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.488958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.492623] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.492814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.492830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.498443] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.576 [2024-11-06 09:08:13.498634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.576 [2024-11-06 09:08:13.498650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.576 [2024-11-06 09:08:13.505997] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.506190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.506206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.509951] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.510136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.510152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.514139] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.514318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.514334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.518307] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.518486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.518503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.524285] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.524520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.524537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.529941] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.530122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.530137] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.533841] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.534021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.534037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.538002] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.538181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.538198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.544320] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.544638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.544655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.554290] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.554610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.554627] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.561508] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.561794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.561810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.568512] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.568693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.568709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.575185] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.575363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.575379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.581535] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.581713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.581737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.586812] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.587011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.587032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.594344] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.594521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.594538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.603475] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.603654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.603671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.609791] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.609988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.610004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.615354] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.615532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.615551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.620238] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.620416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.620438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.627530] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.627910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.627926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.636503] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.636705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.636721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.645828] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.646170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.646187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.657338] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.657656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.657673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.666973] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.667212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.667228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.577 [2024-11-06 09:08:13.678129] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.577 [2024-11-06 09:08:13.678373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.577 [2024-11-06 09:08:13.678390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.838 [2024-11-06 09:08:13.688828] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.838 [2024-11-06 09:08:13.689110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.838 [2024-11-06 09:08:13.689126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.838 [2024-11-06 09:08:13.700265] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.838 [2024-11-06 09:08:13.700477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.838 [2024-11-06 09:08:13.700493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.838 [2024-11-06 09:08:13.711306] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.838 [2024-11-06 09:08:13.711533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.838 [2024-11-06 09:08:13.711549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.838 [2024-11-06 09:08:13.722340] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.838 [2024-11-06 09:08:13.722766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.838 [2024-11-06 09:08:13.722783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.838 [2024-11-06 09:08:13.734136] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.734403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.734422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.745195] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.745465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.745482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.756514] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.756836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.756854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.767445] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.767638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.767654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.777981] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.778206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.778223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.788386] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.788660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.788677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.798751] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.799025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.799041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.809440] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.809708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.809725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.820148] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.820363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.820379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.831021] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.831434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.831451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.842425] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.842667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.842684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.852839] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.853159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.853177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.863958] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.864152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.864169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.873627] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.873914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.873932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.880099] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.880279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.880298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.886934] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.887224] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.887241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.891657] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.891843] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.891864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.899113] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.899291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.899307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.904429] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.904805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.904822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.909009] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.909188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.909203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.913389] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.913568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.913584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.918213] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.918391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.918409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.925980] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.926290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.926307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.931901] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.932136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.932153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.940019] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.940299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.940315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:23.839 [2024-11-06 09:08:13.948396] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:23.839 [2024-11-06 09:08:13.948724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:23.839 [2024-11-06 09:08:13.948741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.100 4478.00 IOPS, 559.75 MiB/s [2024-11-06T08:08:14.213Z] [2024-11-06 09:08:13.957243] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.100 [2024-11-06 09:08:13.957312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.100 [2024-11-06 09:08:13.957327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.100 [2024-11-06 09:08:13.966464] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:13.966520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:0 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:13.966535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:13.973447] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:13.973503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:13.973518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:13.981101] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:13.981375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:13.981391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:13.988689] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:13.988744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:13.988763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:13.993899] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:13.993957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:13.993972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.000090] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.000170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.000185] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.007777] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.007865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.007881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.013869] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.014125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.014141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.021546] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.021627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.021642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.029441] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.029640] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.029655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.037095] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.037152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.037168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.044456] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.044541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.044556] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.053144] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.053202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.053217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.060467] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.060529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.060547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.067843] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.068165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.068180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.074140] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.074194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.074209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.080319] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.080608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.080624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.086984] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.087052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.087067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.093900] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.094120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.094135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.101271] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.101326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.101342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.106887] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.106967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.106982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.112792] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.113034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:0 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.113049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.120652] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.120850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.120866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.128602] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.128661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.128678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.135151] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.135214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.135229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.142224] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.142296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.142311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.148680] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.101 [2024-11-06 09:08:14.148985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.101 [2024-11-06 09:08:14.149001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.101 [2024-11-06 09:08:14.154876] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.102 [2024-11-06 09:08:14.154932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.102 [2024-11-06 09:08:14.154947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.102 [2024-11-06 09:08:14.159487] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.102 [2024-11-06 09:08:14.159547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.102 [2024-11-06 09:08:14.159561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.102 [2024-11-06 09:08:14.168744] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.102 [2024-11-06 09:08:14.168811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.102 [2024-11-06 09:08:14.168827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.102 [2024-11-06 09:08:14.176804] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.102 [2024-11-06 09:08:14.177061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.102 [2024-11-06 09:08:14.177077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.102 [2024-11-06 09:08:14.184938] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.102 [2024-11-06 09:08:14.185182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.102 [2024-11-06 09:08:14.185197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.102 [2024-11-06 09:08:14.191534] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.102 [2024-11-06 09:08:14.191601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.102 [2024-11-06 09:08:14.191616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.102 [2024-11-06 09:08:14.197682] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.102 [2024-11-06 09:08:14.197759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.102 [2024-11-06 09:08:14.197774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.102 [2024-11-06 09:08:14.204148] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.102 [2024-11-06 09:08:14.204286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.102 [2024-11-06 09:08:14.204301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.102 [2024-11-06 09:08:14.210495] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.102 [2024-11-06 09:08:14.210553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.102 [2024-11-06 09:08:14.210568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.363 [2024-11-06 09:08:14.216459] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.363 [2024-11-06 09:08:14.216528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.363 [2024-11-06 09:08:14.216543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.363 [2024-11-06 09:08:14.222515] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.363 [2024-11-06 09:08:14.222578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.363 [2024-11-06 09:08:14.222593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.363 [2024-11-06 09:08:14.230728] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.363 [2024-11-06 09:08:14.230794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.363 [2024-11-06 09:08:14.230810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.363 [2024-11-06 09:08:14.236959] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.363 [2024-11-06 09:08:14.237043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.363 [2024-11-06 09:08:14.237061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.363 [2024-11-06 09:08:14.243156] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.363 [2024-11-06 09:08:14.243210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.363 [2024-11-06 09:08:14.243226] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.363 [2024-11-06 09:08:14.248796] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.363 [2024-11-06 09:08:14.249009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.363 [2024-11-06 09:08:14.249025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.363 [2024-11-06 09:08:14.256526] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.363 [2024-11-06 09:08:14.256603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.363 [2024-11-06 09:08:14.256618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.363 [2024-11-06 09:08:14.264442] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.363 [2024-11-06 09:08:14.264543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.363 [2024-11-06 09:08:14.264558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.270282] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.270349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.270365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.276737] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.276817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.276831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.282812] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.282888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.282903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.287793] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.287848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.287863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.292341] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.292551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.292566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.300068] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.300153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.300168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.305146] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.305337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.305352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.312107] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.312256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.312271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.318833] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.318898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.318913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.325449] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.325507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.325522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.333570] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.333649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.333664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.340575] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.340637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.340652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.346733] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.346813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.346828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.353658] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.353727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.353742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.359350] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.359410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.359425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.365759] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.365822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.365838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.371640] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.371695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.371711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.375896] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.375950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.375965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.381758] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.381901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.381916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.388080] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.388140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.388156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.395081] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.395150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.395165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.399615] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.399671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.399689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.403848] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.403905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.403920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.407753] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.407809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.364 [2024-11-06 09:08:14.407825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.364 [2024-11-06 09:08:14.411652] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.364 [2024-11-06 09:08:14.411704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.365 [2024-11-06 09:08:14.411719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.365 [2024-11-06 09:08:14.416059] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.365 [2024-11-06 09:08:14.416117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.365 [2024-11-06 09:08:14.416133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.365 [2024-11-06 09:08:14.419744] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.365 [2024-11-06 09:08:14.419806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.365 [2024-11-06 09:08:14.419821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.365 [2024-11-06 09:08:14.425821] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.365 [2024-11-06 09:08:14.426143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.365 [2024-11-06 09:08:14.426157] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.365 [2024-11-06 09:08:14.431117] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.365 [2024-11-06 09:08:14.431239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.365 [2024-11-06 09:08:14.431254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.365 [2024-11-06 09:08:14.441512] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.365 [2024-11-06 09:08:14.441623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.365 [2024-11-06 09:08:14.441638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.365 [2024-11-06 09:08:14.451576] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.365 [2024-11-06 09:08:14.451876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.365 [2024-11-06 09:08:14.451892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.365 [2024-11-06 09:08:14.462564] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.365 [2024-11-06 09:08:14.462836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.365 [2024-11-06 09:08:14.462852] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.365 [2024-11-06 09:08:14.472936] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.365 [2024-11-06 09:08:14.473272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.365 [2024-11-06 09:08:14.473288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.483606] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.483929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.483945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.494783] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.495015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.495030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.505212] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.505564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.505579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.516755] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.517084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.517100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.526679] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.526803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.526818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.531397] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.531450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.531468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.535569] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.535641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.535656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.540848] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.540904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.540920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.545709] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.545780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.545795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.551440] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.551515] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.551530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.555804] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.555870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.555886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.561205] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.561358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.561373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.568626] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.568740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.568760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.573334] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.573394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.573410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.579046] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.579121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.579136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.583389] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.583461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.583476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.588160] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.588444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.588460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.592532] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.592584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.592599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.598398] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.598458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.598473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.603367] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.603421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.603436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.627 [2024-11-06 09:08:14.607484] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.627 [2024-11-06 09:08:14.607588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.627 [2024-11-06 09:08:14.607603] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.613142] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.613207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.613223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.621253] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.621553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.621569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.627800] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.628065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.628081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.634885] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.634961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.634976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.641389] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.641467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.641483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.645966] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.646034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.646048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.650293] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.650349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.650364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.654539] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.654591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.654606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.660423] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.660479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.660494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.664277] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.664334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.664349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.668472] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.668765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.668784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.675983] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.676038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.676053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.683506] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.683825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.683841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.690357] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.690413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.690428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.696190] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.696276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.696291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.704774] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.705131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.705148] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.712266] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.712321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.712336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.718820] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.718886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.718901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.726321] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.726375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2496 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.726390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.628 [2024-11-06 09:08:14.733612] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.628 [2024-11-06 09:08:14.733955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.628 [2024-11-06 09:08:14.733971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.739821] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.739887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.739902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.747448] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.747519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.747534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.755500] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.755569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1920 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.755584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.762218] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.762286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.762301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.766798] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.766886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.766901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.771452] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.771510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.771525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.776699] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.777013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.777029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.781403] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.781460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.781476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.785823] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.785877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.785892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.790413] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.790467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.790482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.794989] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.795111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.795126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.804795] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.804858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.804873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.812911] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.813214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.813230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.822078] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.822155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.822170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.826851] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.827112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.827127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.833105] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.833341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.833356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.838930] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.839177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.839194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.844924] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.844989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.845004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.851586] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.851671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.851686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.855633] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.855683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.855699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.859688] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.859758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.859774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.863764] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.863822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.863836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.868946] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.869012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.891 [2024-11-06 09:08:14.869027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.891 [2024-11-06 09:08:14.874184] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.891 [2024-11-06 09:08:14.874241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.874255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.877890] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.877942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.877957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.882210] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.882308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.882323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.887534] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.887616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.887631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.891673] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.891742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.891764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.896248] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.896312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.896327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.904137] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.904192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.904207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.908320] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.908384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.908399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.912296] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.912377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.912392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.916232] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.916299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.916314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.920098] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.920221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.920236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.924974] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.925043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.925058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.930269] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.930353] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.930368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.937042] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.937117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.937132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.941315] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.941370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.941385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.945703] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.945762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.945777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.949871] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.949944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.949959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:32:24.892 [2024-11-06 09:08:14.954438] tcp.c:2233:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x8fb390) with pdu=0x200016efef90 00:32:24.892 [2024-11-06 09:08:14.954494] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:24.892 [2024-11-06 09:08:14.954509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:32:24.892 4713.00 IOPS, 589.12 MiB/s 00:32:24.892 Latency(us) 00:32:24.892 [2024-11-06T08:08:15.005Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:24.892 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:32:24.892 nvme0n1 : 2.00 4716.46 589.56 0.00 0.00 3388.31 1283.41 13216.43 00:32:24.892 [2024-11-06T08:08:15.005Z] =================================================================================================================== 00:32:24.892 [2024-11-06T08:08:15.005Z] Total : 4716.46 589.56 0.00 0.00 3388.31 1283.41 13216.43 00:32:24.892 { 00:32:24.892 "results": [ 00:32:24.892 { 00:32:24.892 "job": "nvme0n1", 00:32:24.892 "core_mask": "0x2", 00:32:24.892 "workload": "randwrite", 00:32:24.892 "status": "finished", 00:32:24.892 "queue_depth": 16, 00:32:24.892 "io_size": 131072, 00:32:24.892 "runtime": 2.002984, 00:32:24.892 "iops": 4716.463037148575, 00:32:24.892 "mibps": 589.5578796435718, 00:32:24.892 "io_failed": 0, 00:32:24.892 "io_timeout": 0, 00:32:24.892 "avg_latency_us": 3388.308364560178, 00:32:24.892 "min_latency_us": 1283.4133333333334, 00:32:24.892 "max_latency_us": 13216.426666666666 00:32:24.892 } 00:32:24.892 ], 00:32:24.892 "core_count": 1 00:32:24.892 } 00:32:24.892 09:08:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:32:24.892 09:08:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:32:24.892 09:08:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:32:24.892 | .driver_specific 00:32:24.892 | .nvme_error 00:32:24.892 | .status_code 00:32:24.892 | .command_transient_transport_error' 00:32:24.892 09:08:14 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 304 > 0 )) 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 2067510 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@950 -- # '[' -z 2067510 ']' 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # kill -0 2067510 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # uname 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2067510 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2067510' 00:32:25.153 killing process with pid 2067510 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@969 -- # kill 2067510 00:32:25.153 Received shutdown signal, test time was about 2.000000 seconds 00:32:25.153 00:32:25.153 Latency(us) 00:32:25.153 [2024-11-06T08:08:15.266Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:25.153 [2024-11-06T08:08:15.266Z] =================================================================================================================== 00:32:25.153 [2024-11-06T08:08:15.266Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:32:25.153 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@974 -- # wait 2067510 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- host/digest.sh@116 -- # killprocess 2065108 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@950 -- # '[' -z 2065108 ']' 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # kill -0 2065108 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # uname 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2065108 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2065108' 00:32:25.413 killing process with pid 2065108 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@969 -- # kill 2065108 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@974 -- # wait 2065108 00:32:25.413 00:32:25.413 real 0m16.475s 00:32:25.413 user 0m32.589s 00:32:25.413 sys 0m3.519s 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:25.413 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:32:25.413 ************************************ 00:32:25.413 END TEST nvmf_digest_error 00:32:25.413 ************************************ 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@149 -- # trap - SIGINT SIGTERM EXIT 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- host/digest.sh@150 -- # nvmftestfini 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@514 -- # nvmfcleanup 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@121 -- # sync 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@124 -- # set +e 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:25.673 rmmod nvme_tcp 00:32:25.673 rmmod nvme_fabrics 00:32:25.673 rmmod nvme_keyring 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@128 -- # set -e 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@129 -- # return 0 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@515 -- # '[' -n 2065108 ']' 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@516 -- # killprocess 2065108 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@950 -- # '[' -z 2065108 ']' 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@954 -- # kill -0 2065108 00:32:25.673 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (2065108) - No such process 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@977 -- # echo 'Process with pid 2065108 is not found' 00:32:25.673 Process with pid 2065108 is not found 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@297 -- # iptr 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@789 -- # iptables-save 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@789 -- # iptables-restore 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:25.673 09:08:15 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:27.589 09:08:17 nvmf_tcp.nvmf_host.nvmf_digest -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:27.851 00:32:27.851 real 0m42.274s 00:32:27.851 user 1m6.174s 00:32:27.851 sys 0m12.604s 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:32:27.851 ************************************ 00:32:27.851 END TEST nvmf_digest 00:32:27.851 ************************************ 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@36 -- # [[ 0 -eq 1 ]] 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@41 -- # [[ 0 -eq 1 ]] 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@46 -- # [[ phy == phy ]] 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@47 -- # run_test nvmf_bdevperf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:32:27.851 ************************************ 00:32:27.851 START TEST nvmf_bdevperf 00:32:27.851 ************************************ 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:32:27.851 * Looking for test storage... 00:32:27.851 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1689 -- # lcov --version 00:32:27.851 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@336 -- # IFS=.-: 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@336 -- # read -ra ver1 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@337 -- # IFS=.-: 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@337 -- # read -ra ver2 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@338 -- # local 'op=<' 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@340 -- # ver1_l=2 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@341 -- # ver2_l=1 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@344 -- # case "$op" in 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@345 -- # : 1 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@365 -- # decimal 1 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@353 -- # local d=1 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@355 -- # echo 1 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@365 -- # ver1[v]=1 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@366 -- # decimal 2 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@353 -- # local d=2 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@355 -- # echo 2 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@366 -- # ver2[v]=2 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@368 -- # return 0 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:32:28.114 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:28.114 --rc genhtml_branch_coverage=1 00:32:28.114 --rc genhtml_function_coverage=1 00:32:28.114 --rc genhtml_legend=1 00:32:28.114 --rc geninfo_all_blocks=1 00:32:28.114 --rc geninfo_unexecuted_blocks=1 00:32:28.114 00:32:28.114 ' 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:32:28.114 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:28.114 --rc genhtml_branch_coverage=1 00:32:28.114 --rc genhtml_function_coverage=1 00:32:28.114 --rc genhtml_legend=1 00:32:28.114 --rc geninfo_all_blocks=1 00:32:28.114 --rc geninfo_unexecuted_blocks=1 00:32:28.114 00:32:28.114 ' 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:32:28.114 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:28.114 --rc genhtml_branch_coverage=1 00:32:28.114 --rc genhtml_function_coverage=1 00:32:28.114 --rc genhtml_legend=1 00:32:28.114 --rc geninfo_all_blocks=1 00:32:28.114 --rc geninfo_unexecuted_blocks=1 00:32:28.114 00:32:28.114 ' 00:32:28.114 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:32:28.114 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:28.114 --rc genhtml_branch_coverage=1 00:32:28.114 --rc genhtml_function_coverage=1 00:32:28.115 --rc genhtml_legend=1 00:32:28.115 --rc geninfo_all_blocks=1 00:32:28.115 --rc geninfo_unexecuted_blocks=1 00:32:28.115 00:32:28.115 ' 00:32:28.115 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:28.115 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@7 -- # uname -s 00:32:28.115 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:28.115 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:28.115 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:28.115 09:08:17 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@15 -- # shopt -s extglob 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@5 -- # export PATH 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@51 -- # : 0 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:32:28.115 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@11 -- # MALLOC_BDEV_SIZE=64 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@24 -- # nvmftestinit 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@474 -- # prepare_net_devs 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@436 -- # local -g is_hw=no 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@438 -- # remove_spdk_ns 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@309 -- # xtrace_disable 00:32:28.115 09:08:18 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@315 -- # pci_devs=() 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@315 -- # local -a pci_devs 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@316 -- # pci_net_devs=() 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@317 -- # pci_drivers=() 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@317 -- # local -A pci_drivers 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@319 -- # net_devs=() 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@319 -- # local -ga net_devs 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@320 -- # e810=() 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@320 -- # local -ga e810 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@321 -- # x722=() 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@321 -- # local -ga x722 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@322 -- # mlx=() 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@322 -- # local -ga mlx 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:32:36.260 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:32:36.260 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:32:36.260 Found net devices under 0000:4b:00.0: cvl_0_0 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@416 -- # [[ up == up ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:32:36.260 Found net devices under 0000:4b:00.1: cvl_0_1 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@440 -- # is_hw=yes 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:32:36.260 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:32:36.261 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:32:36.261 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:32:36.261 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:32:36.261 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:32:36.261 09:08:24 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:32:36.261 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:32:36.261 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.656 ms 00:32:36.261 00:32:36.261 --- 10.0.0.2 ping statistics --- 00:32:36.261 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:36.261 rtt min/avg/max/mdev = 0.656/0.656/0.656/0.000 ms 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:32:36.261 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:32:36.261 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.288 ms 00:32:36.261 00:32:36.261 --- 10.0.0.1 ping statistics --- 00:32:36.261 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:32:36.261 rtt min/avg/max/mdev = 0.288/0.288/0.288/0.000 ms 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@448 -- # return 0 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@25 -- # tgt_init 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@724 -- # xtrace_disable 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@507 -- # nvmfpid=2072404 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@508 -- # waitforlisten 2072404 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@831 -- # '[' -z 2072404 ']' 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:36.261 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:36.261 09:08:25 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:36.261 [2024-11-06 09:08:25.263405] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:36.261 [2024-11-06 09:08:25.263473] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:36.261 [2024-11-06 09:08:25.363020] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:32:36.261 [2024-11-06 09:08:25.403233] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:36.261 [2024-11-06 09:08:25.403273] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:36.261 [2024-11-06 09:08:25.403282] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:36.261 [2024-11-06 09:08:25.403288] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:36.261 [2024-11-06 09:08:25.403295] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:36.261 [2024-11-06 09:08:25.404791] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:32:36.261 [2024-11-06 09:08:25.404957] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:36.261 [2024-11-06 09:08:25.404958] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@864 -- # return 0 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@730 -- # xtrace_disable 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:36.261 [2024-11-06 09:08:26.107899] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:36.261 Malloc0 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:36.261 [2024-11-06 09:08:26.175947] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 128 -o 4096 -w verify -t 1 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@27 -- # gen_nvmf_target_json 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@558 -- # config=() 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@558 -- # local subsystem config 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:32:36.261 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:32:36.261 { 00:32:36.261 "params": { 00:32:36.262 "name": "Nvme$subsystem", 00:32:36.262 "trtype": "$TEST_TRANSPORT", 00:32:36.262 "traddr": "$NVMF_FIRST_TARGET_IP", 00:32:36.262 "adrfam": "ipv4", 00:32:36.262 "trsvcid": "$NVMF_PORT", 00:32:36.262 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:32:36.262 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:32:36.262 "hdgst": ${hdgst:-false}, 00:32:36.262 "ddgst": ${ddgst:-false} 00:32:36.262 }, 00:32:36.262 "method": "bdev_nvme_attach_controller" 00:32:36.262 } 00:32:36.262 EOF 00:32:36.262 )") 00:32:36.262 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@580 -- # cat 00:32:36.262 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # jq . 00:32:36.262 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@583 -- # IFS=, 00:32:36.262 09:08:26 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:32:36.262 "params": { 00:32:36.262 "name": "Nvme1", 00:32:36.262 "trtype": "tcp", 00:32:36.262 "traddr": "10.0.0.2", 00:32:36.262 "adrfam": "ipv4", 00:32:36.262 "trsvcid": "4420", 00:32:36.262 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:36.262 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:32:36.262 "hdgst": false, 00:32:36.262 "ddgst": false 00:32:36.262 }, 00:32:36.262 "method": "bdev_nvme_attach_controller" 00:32:36.262 }' 00:32:36.262 [2024-11-06 09:08:26.230472] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:36.262 [2024-11-06 09:08:26.230521] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2072616 ] 00:32:36.262 [2024-11-06 09:08:26.300293] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:36.262 [2024-11-06 09:08:26.336202] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:36.523 Running I/O for 1 seconds... 00:32:37.907 8938.00 IOPS, 34.91 MiB/s 00:32:37.907 Latency(us) 00:32:37.907 [2024-11-06T08:08:28.020Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:37.907 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:32:37.907 Verification LBA range: start 0x0 length 0x4000 00:32:37.907 Nvme1n1 : 1.01 8942.91 34.93 0.00 0.00 14254.91 2962.77 14964.05 00:32:37.907 [2024-11-06T08:08:28.020Z] =================================================================================================================== 00:32:37.907 [2024-11-06T08:08:28.020Z] Total : 8942.91 34.93 0.00 0.00 14254.91 2962.77 14964.05 00:32:37.907 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@30 -- # bdevperfpid=2072948 00:32:37.907 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@32 -- # sleep 3 00:32:37.907 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -q 128 -o 4096 -w verify -t 15 -f 00:32:37.907 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@29 -- # gen_nvmf_target_json 00:32:37.907 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@558 -- # config=() 00:32:37.908 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@558 -- # local subsystem config 00:32:37.908 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:32:37.908 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:32:37.908 { 00:32:37.908 "params": { 00:32:37.908 "name": "Nvme$subsystem", 00:32:37.908 "trtype": "$TEST_TRANSPORT", 00:32:37.908 "traddr": "$NVMF_FIRST_TARGET_IP", 00:32:37.908 "adrfam": "ipv4", 00:32:37.908 "trsvcid": "$NVMF_PORT", 00:32:37.908 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:32:37.908 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:32:37.908 "hdgst": ${hdgst:-false}, 00:32:37.908 "ddgst": ${ddgst:-false} 00:32:37.908 }, 00:32:37.908 "method": "bdev_nvme_attach_controller" 00:32:37.908 } 00:32:37.908 EOF 00:32:37.908 )") 00:32:37.908 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@580 -- # cat 00:32:37.908 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@582 -- # jq . 00:32:37.908 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@583 -- # IFS=, 00:32:37.908 09:08:27 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:32:37.908 "params": { 00:32:37.908 "name": "Nvme1", 00:32:37.908 "trtype": "tcp", 00:32:37.908 "traddr": "10.0.0.2", 00:32:37.908 "adrfam": "ipv4", 00:32:37.908 "trsvcid": "4420", 00:32:37.908 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:32:37.908 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:32:37.908 "hdgst": false, 00:32:37.908 "ddgst": false 00:32:37.908 }, 00:32:37.908 "method": "bdev_nvme_attach_controller" 00:32:37.908 }' 00:32:37.908 [2024-11-06 09:08:27.794838] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:37.908 [2024-11-06 09:08:27.794892] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2072948 ] 00:32:37.908 [2024-11-06 09:08:27.865318] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:32:37.908 [2024-11-06 09:08:27.900586] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:32:38.168 Running I/O for 15 seconds... 00:32:40.492 11061.00 IOPS, 43.21 MiB/s [2024-11-06T08:08:30.869Z] 11079.00 IOPS, 43.28 MiB/s [2024-11-06T08:08:30.869Z] 09:08:30 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@33 -- # kill -9 2072404 00:32:40.756 09:08:30 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@35 -- # sleep 3 00:32:40.756 [2024-11-06 09:08:30.759303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:96856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:96864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:96872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:96880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:96888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:96896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:96904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:96912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:96920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:96928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:96936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:96944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:96952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:96960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:96968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:96976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:96984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:96992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:97000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:97008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:97016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:97024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:97032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:97040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:97048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:97056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:97064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:97072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:97080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:97088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:97096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:97104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:97112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.756 [2024-11-06 09:08:30.759968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:97120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.756 [2024-11-06 09:08:30.759975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.759984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:97128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.757 [2024-11-06 09:08:30.759991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:97136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.757 [2024-11-06 09:08:30.760008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:97144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.757 [2024-11-06 09:08:30.760025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:97152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.757 [2024-11-06 09:08:30.760042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:97160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.757 [2024-11-06 09:08:30.760058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:97168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.757 [2024-11-06 09:08:30.760075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:97176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.757 [2024-11-06 09:08:30.760092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:97240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:97248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:97256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:97264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:97272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:97280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:97288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:97296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:97304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:97312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:97320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:97328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:97336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760309] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:97344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:97352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760357] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:97360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760374] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:97368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:97376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:97384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760415] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:97392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:97400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:97408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:97416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:97424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:97432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:97440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760540] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:97448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760559] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:97456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:97464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:97472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:97480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:97488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.757 [2024-11-06 09:08:30.760642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:97496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.757 [2024-11-06 09:08:30.760649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:97504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:97512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:97520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:97528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:97536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:97544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:97552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:97560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:97568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:97576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:97584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:97592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760864] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:97600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:97608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:97616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:97624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:97632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760937] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:97640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:97648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:97656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.760988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.760998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:97664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:97672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:97680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:97688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:97696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:97704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:97712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:97720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:97728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:97736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:97744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761181] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:97752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761199] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:97760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:97768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:97776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:97784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:97792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:97800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:97808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.758 [2024-11-06 09:08:30.761306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.758 [2024-11-06 09:08:30.761315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:97816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.759 [2024-11-06 09:08:30.761322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:97824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.759 [2024-11-06 09:08:30.761339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:97832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.759 [2024-11-06 09:08:30.761356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:97840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.759 [2024-11-06 09:08:30.761372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:97848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.759 [2024-11-06 09:08:30.761388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:97856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.759 [2024-11-06 09:08:30.761407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:97864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.759 [2024-11-06 09:08:30.761423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:97872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:32:40.759 [2024-11-06 09:08:30.761439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:97184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.759 [2024-11-06 09:08:30.761456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:97192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.759 [2024-11-06 09:08:30.761473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:97200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.759 [2024-11-06 09:08:30.761490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:97208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.759 [2024-11-06 09:08:30.761508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:97216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.759 [2024-11-06 09:08:30.761524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:97224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:32:40.759 [2024-11-06 09:08:30.761541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.761550] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe37000 is same with the state(6) to be set 00:32:40.759 [2024-11-06 09:08:30.761559] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:32:40.759 [2024-11-06 09:08:30.761566] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:32:40.759 [2024-11-06 09:08:30.761572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:97232 len:8 PRP1 0x0 PRP2 0x0 00:32:40.759 [2024-11-06 09:08:30.761580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:32:40.759 [2024-11-06 09:08:30.765178] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:40.759 [2024-11-06 09:08:30.765231] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:40.759 [2024-11-06 09:08:30.766130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:40.759 [2024-11-06 09:08:30.766169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:40.759 [2024-11-06 09:08:30.766180] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:40.759 [2024-11-06 09:08:30.766425] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:40.759 [2024-11-06 09:08:30.766648] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:40.759 [2024-11-06 09:08:30.766657] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:40.759 [2024-11-06 09:08:30.766666] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:40.759 [2024-11-06 09:08:30.770202] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:40.759 [2024-11-06 09:08:30.779359] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:40.759 [2024-11-06 09:08:30.780038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:40.759 [2024-11-06 09:08:30.780077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:40.759 [2024-11-06 09:08:30.780088] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:40.759 [2024-11-06 09:08:30.780325] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:40.759 [2024-11-06 09:08:30.780547] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:40.759 [2024-11-06 09:08:30.780557] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:40.759 [2024-11-06 09:08:30.780565] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:40.759 [2024-11-06 09:08:30.784094] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:40.759 [2024-11-06 09:08:30.793221] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:40.759 [2024-11-06 09:08:30.793959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:40.759 [2024-11-06 09:08:30.793997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:40.759 [2024-11-06 09:08:30.794008] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:40.759 [2024-11-06 09:08:30.794245] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:40.759 [2024-11-06 09:08:30.794467] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:40.759 [2024-11-06 09:08:30.794476] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:40.759 [2024-11-06 09:08:30.794483] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:40.759 [2024-11-06 09:08:30.798015] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:40.759 [2024-11-06 09:08:30.807158] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:40.759 [2024-11-06 09:08:30.807827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:40.759 [2024-11-06 09:08:30.807865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:40.759 [2024-11-06 09:08:30.807878] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:40.759 [2024-11-06 09:08:30.808119] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:40.759 [2024-11-06 09:08:30.808341] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:40.759 [2024-11-06 09:08:30.808355] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:40.759 [2024-11-06 09:08:30.808363] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:40.759 [2024-11-06 09:08:30.811891] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:40.759 [2024-11-06 09:08:30.821019] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:40.759 [2024-11-06 09:08:30.821547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:40.759 [2024-11-06 09:08:30.821566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:40.759 [2024-11-06 09:08:30.821575] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:40.759 [2024-11-06 09:08:30.821798] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:40.759 [2024-11-06 09:08:30.822016] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:40.759 [2024-11-06 09:08:30.822030] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:40.759 [2024-11-06 09:08:30.822037] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:40.759 [2024-11-06 09:08:30.825552] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:40.759 [2024-11-06 09:08:30.834887] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:40.759 [2024-11-06 09:08:30.835462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:40.759 [2024-11-06 09:08:30.835478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:40.759 [2024-11-06 09:08:30.835486] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:40.759 [2024-11-06 09:08:30.835703] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:40.759 [2024-11-06 09:08:30.835926] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:40.759 [2024-11-06 09:08:30.835935] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:40.759 [2024-11-06 09:08:30.835942] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:40.759 [2024-11-06 09:08:30.839457] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:40.760 [2024-11-06 09:08:30.848790] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:40.760 [2024-11-06 09:08:30.849344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:40.760 [2024-11-06 09:08:30.849360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:40.760 [2024-11-06 09:08:30.849367] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:40.760 [2024-11-06 09:08:30.849584] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:40.760 [2024-11-06 09:08:30.849808] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:40.760 [2024-11-06 09:08:30.849816] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:40.760 [2024-11-06 09:08:30.849823] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:40.760 [2024-11-06 09:08:30.853343] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:40.760 [2024-11-06 09:08:30.862676] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:40.760 [2024-11-06 09:08:30.863206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:40.760 [2024-11-06 09:08:30.863222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:40.760 [2024-11-06 09:08:30.863229] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:40.760 [2024-11-06 09:08:30.863447] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:40.760 [2024-11-06 09:08:30.863663] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:40.760 [2024-11-06 09:08:30.863672] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:40.760 [2024-11-06 09:08:30.863679] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.021 [2024-11-06 09:08:30.867210] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.021 [2024-11-06 09:08:30.876554] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.021 [2024-11-06 09:08:30.877097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.021 [2024-11-06 09:08:30.877113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.021 [2024-11-06 09:08:30.877121] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.021 [2024-11-06 09:08:30.877339] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.021 [2024-11-06 09:08:30.877555] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.021 [2024-11-06 09:08:30.877564] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.021 [2024-11-06 09:08:30.877571] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.021 [2024-11-06 09:08:30.881098] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.021 [2024-11-06 09:08:30.890432] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.021 [2024-11-06 09:08:30.891038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.021 [2024-11-06 09:08:30.891077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.021 [2024-11-06 09:08:30.891088] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.021 [2024-11-06 09:08:30.891325] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.022 [2024-11-06 09:08:30.891546] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.022 [2024-11-06 09:08:30.891555] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.022 [2024-11-06 09:08:30.891563] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.022 [2024-11-06 09:08:30.895092] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.022 [2024-11-06 09:08:30.904305] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.022 [2024-11-06 09:08:30.904856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.022 [2024-11-06 09:08:30.904895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.022 [2024-11-06 09:08:30.904907] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.022 [2024-11-06 09:08:30.905147] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.022 [2024-11-06 09:08:30.905369] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.022 [2024-11-06 09:08:30.905385] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.022 [2024-11-06 09:08:30.905393] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.022 [2024-11-06 09:08:30.908920] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.022 [2024-11-06 09:08:30.918263] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.022 [2024-11-06 09:08:30.918954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.022 [2024-11-06 09:08:30.918992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.022 [2024-11-06 09:08:30.919003] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.022 [2024-11-06 09:08:30.919239] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.022 [2024-11-06 09:08:30.919461] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.022 [2024-11-06 09:08:30.919470] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.022 [2024-11-06 09:08:30.919478] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.022 [2024-11-06 09:08:30.923005] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.022 [2024-11-06 09:08:30.932130] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.022 [2024-11-06 09:08:30.932817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.022 [2024-11-06 09:08:30.932854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.022 [2024-11-06 09:08:30.932865] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.022 [2024-11-06 09:08:30.933102] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.022 [2024-11-06 09:08:30.933324] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.022 [2024-11-06 09:08:30.933333] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.022 [2024-11-06 09:08:30.933341] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.022 [2024-11-06 09:08:30.936871] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.022 [2024-11-06 09:08:30.945990] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.022 [2024-11-06 09:08:30.946667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.022 [2024-11-06 09:08:30.946705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.022 [2024-11-06 09:08:30.946716] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.022 [2024-11-06 09:08:30.946966] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.022 [2024-11-06 09:08:30.947189] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.022 [2024-11-06 09:08:30.947198] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.022 [2024-11-06 09:08:30.947206] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.022 [2024-11-06 09:08:30.950726] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.022 [2024-11-06 09:08:30.959857] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.022 [2024-11-06 09:08:30.960504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.022 [2024-11-06 09:08:30.960541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.022 [2024-11-06 09:08:30.960552] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.022 [2024-11-06 09:08:30.960798] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.022 [2024-11-06 09:08:30.961021] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.022 [2024-11-06 09:08:30.961030] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.022 [2024-11-06 09:08:30.961038] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.022 [2024-11-06 09:08:30.964558] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.022 [2024-11-06 09:08:30.973684] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.022 [2024-11-06 09:08:30.974361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.022 [2024-11-06 09:08:30.974399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.022 [2024-11-06 09:08:30.974410] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.022 [2024-11-06 09:08:30.974646] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.022 [2024-11-06 09:08:30.974878] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.022 [2024-11-06 09:08:30.974888] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.022 [2024-11-06 09:08:30.974895] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.022 [2024-11-06 09:08:30.978426] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.022 [2024-11-06 09:08:30.987557] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.022 [2024-11-06 09:08:30.988190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.022 [2024-11-06 09:08:30.988228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.022 [2024-11-06 09:08:30.988239] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.022 [2024-11-06 09:08:30.988476] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.022 [2024-11-06 09:08:30.988698] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.022 [2024-11-06 09:08:30.988712] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.022 [2024-11-06 09:08:30.988719] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.022 [2024-11-06 09:08:30.992251] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.022 [2024-11-06 09:08:31.001383] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.022 [2024-11-06 09:08:31.002061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.022 [2024-11-06 09:08:31.002099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.022 [2024-11-06 09:08:31.002110] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.022 [2024-11-06 09:08:31.002346] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.022 [2024-11-06 09:08:31.002568] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.022 [2024-11-06 09:08:31.002577] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.022 [2024-11-06 09:08:31.002585] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.022 [2024-11-06 09:08:31.006114] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.022 [2024-11-06 09:08:31.015236] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.022 [2024-11-06 09:08:31.015848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.022 [2024-11-06 09:08:31.015886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.022 [2024-11-06 09:08:31.015899] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.022 [2024-11-06 09:08:31.016140] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.022 [2024-11-06 09:08:31.016362] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.022 [2024-11-06 09:08:31.016378] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.022 [2024-11-06 09:08:31.016386] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.022 [2024-11-06 09:08:31.019918] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.022 [2024-11-06 09:08:31.029044] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.022 [2024-11-06 09:08:31.029585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.022 [2024-11-06 09:08:31.029604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.022 [2024-11-06 09:08:31.029613] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.023 [2024-11-06 09:08:31.029837] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.023 [2024-11-06 09:08:31.030055] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.023 [2024-11-06 09:08:31.030064] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.023 [2024-11-06 09:08:31.030072] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.023 [2024-11-06 09:08:31.033589] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.023 [2024-11-06 09:08:31.042919] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.023 [2024-11-06 09:08:31.043441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.023 [2024-11-06 09:08:31.043479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.023 [2024-11-06 09:08:31.043490] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.023 [2024-11-06 09:08:31.043727] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.023 [2024-11-06 09:08:31.043957] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.023 [2024-11-06 09:08:31.043967] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.023 [2024-11-06 09:08:31.043974] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.023 [2024-11-06 09:08:31.047498] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.023 [2024-11-06 09:08:31.056837] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.023 [2024-11-06 09:08:31.057469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.023 [2024-11-06 09:08:31.057507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.023 [2024-11-06 09:08:31.057517] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.023 [2024-11-06 09:08:31.057765] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.023 [2024-11-06 09:08:31.057987] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.023 [2024-11-06 09:08:31.057997] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.023 [2024-11-06 09:08:31.058004] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.023 [2024-11-06 09:08:31.061525] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.023 [2024-11-06 09:08:31.070649] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.023 [2024-11-06 09:08:31.071286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.023 [2024-11-06 09:08:31.071323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.023 [2024-11-06 09:08:31.071334] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.023 [2024-11-06 09:08:31.071571] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.023 [2024-11-06 09:08:31.071802] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.023 [2024-11-06 09:08:31.071812] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.023 [2024-11-06 09:08:31.071820] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.023 [2024-11-06 09:08:31.075354] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.023 [2024-11-06 09:08:31.084498] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.023 [2024-11-06 09:08:31.085032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.023 [2024-11-06 09:08:31.085079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.023 [2024-11-06 09:08:31.085090] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.023 [2024-11-06 09:08:31.085327] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.023 [2024-11-06 09:08:31.085549] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.023 [2024-11-06 09:08:31.085557] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.023 [2024-11-06 09:08:31.085565] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.023 [2024-11-06 09:08:31.089100] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.023 [2024-11-06 09:08:31.098311] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.023 [2024-11-06 09:08:31.098856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.023 [2024-11-06 09:08:31.098894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.023 [2024-11-06 09:08:31.098906] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.023 [2024-11-06 09:08:31.099147] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.023 [2024-11-06 09:08:31.099369] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.023 [2024-11-06 09:08:31.099378] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.023 [2024-11-06 09:08:31.099386] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.023 [2024-11-06 09:08:31.102925] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.023 [2024-11-06 09:08:31.112263] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.023 [2024-11-06 09:08:31.112862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.023 [2024-11-06 09:08:31.112900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.023 [2024-11-06 09:08:31.112911] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.023 [2024-11-06 09:08:31.113148] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.023 [2024-11-06 09:08:31.113370] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.023 [2024-11-06 09:08:31.113379] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.023 [2024-11-06 09:08:31.113387] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.023 [2024-11-06 09:08:31.116917] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.023 [2024-11-06 09:08:31.126090] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.023 [2024-11-06 09:08:31.126675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.023 [2024-11-06 09:08:31.126693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.023 [2024-11-06 09:08:31.126700] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.023 [2024-11-06 09:08:31.126930] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.023 [2024-11-06 09:08:31.127148] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.023 [2024-11-06 09:08:31.127157] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.023 [2024-11-06 09:08:31.127164] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.023 [2024-11-06 09:08:31.130679] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.285 [2024-11-06 09:08:31.140009] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.285 [2024-11-06 09:08:31.140540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.285 [2024-11-06 09:08:31.140556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.285 [2024-11-06 09:08:31.140564] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.285 [2024-11-06 09:08:31.140788] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.285 [2024-11-06 09:08:31.141006] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.285 [2024-11-06 09:08:31.141014] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.285 [2024-11-06 09:08:31.141021] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.285 [2024-11-06 09:08:31.144533] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.285 [2024-11-06 09:08:31.153862] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.285 [2024-11-06 09:08:31.154389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.285 [2024-11-06 09:08:31.154404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.285 [2024-11-06 09:08:31.154412] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.285 [2024-11-06 09:08:31.154629] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.285 [2024-11-06 09:08:31.154851] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.285 [2024-11-06 09:08:31.154860] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.285 [2024-11-06 09:08:31.154867] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.285 [2024-11-06 09:08:31.158381] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.285 [2024-11-06 09:08:31.167703] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.285 [2024-11-06 09:08:31.168237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.285 [2024-11-06 09:08:31.168252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.285 [2024-11-06 09:08:31.168260] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.285 [2024-11-06 09:08:31.168476] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.285 [2024-11-06 09:08:31.168693] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.285 [2024-11-06 09:08:31.168705] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.285 [2024-11-06 09:08:31.168712] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.285 [2024-11-06 09:08:31.172234] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.285 9497.00 IOPS, 37.10 MiB/s [2024-11-06T08:08:31.398Z] [2024-11-06 09:08:31.183227] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.285 [2024-11-06 09:08:31.183859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.285 [2024-11-06 09:08:31.183898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.285 [2024-11-06 09:08:31.183910] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.285 [2024-11-06 09:08:31.184151] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.285 [2024-11-06 09:08:31.184372] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.285 [2024-11-06 09:08:31.184381] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.285 [2024-11-06 09:08:31.184389] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.285 [2024-11-06 09:08:31.187921] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.285 [2024-11-06 09:08:31.197047] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.285 [2024-11-06 09:08:31.197718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.285 [2024-11-06 09:08:31.197762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.285 [2024-11-06 09:08:31.197775] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.285 [2024-11-06 09:08:31.198014] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.285 [2024-11-06 09:08:31.198235] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.285 [2024-11-06 09:08:31.198244] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.285 [2024-11-06 09:08:31.198252] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.285 [2024-11-06 09:08:31.201789] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.285 [2024-11-06 09:08:31.210914] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.285 [2024-11-06 09:08:31.211540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.285 [2024-11-06 09:08:31.211578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.285 [2024-11-06 09:08:31.211589] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.285 [2024-11-06 09:08:31.211836] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.286 [2024-11-06 09:08:31.212058] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.286 [2024-11-06 09:08:31.212067] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.286 [2024-11-06 09:08:31.212075] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.286 [2024-11-06 09:08:31.215602] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.286 [2024-11-06 09:08:31.224727] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.286 [2024-11-06 09:08:31.225405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.286 [2024-11-06 09:08:31.225443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.286 [2024-11-06 09:08:31.225454] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.286 [2024-11-06 09:08:31.225692] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.286 [2024-11-06 09:08:31.225923] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.286 [2024-11-06 09:08:31.225933] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.286 [2024-11-06 09:08:31.225941] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.286 [2024-11-06 09:08:31.229463] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.286 [2024-11-06 09:08:31.238590] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.286 [2024-11-06 09:08:31.239234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.286 [2024-11-06 09:08:31.239272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.286 [2024-11-06 09:08:31.239283] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.286 [2024-11-06 09:08:31.239520] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.286 [2024-11-06 09:08:31.239742] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.286 [2024-11-06 09:08:31.239760] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.286 [2024-11-06 09:08:31.239768] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.286 [2024-11-06 09:08:31.243291] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.286 [2024-11-06 09:08:31.252426] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.286 [2024-11-06 09:08:31.252961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.286 [2024-11-06 09:08:31.252981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.286 [2024-11-06 09:08:31.252989] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.286 [2024-11-06 09:08:31.253208] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.286 [2024-11-06 09:08:31.253425] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.286 [2024-11-06 09:08:31.253434] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.286 [2024-11-06 09:08:31.253441] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.286 [2024-11-06 09:08:31.256959] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.286 [2024-11-06 09:08:31.266283] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.286 [2024-11-06 09:08:31.266821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.286 [2024-11-06 09:08:31.266842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.286 [2024-11-06 09:08:31.266850] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.286 [2024-11-06 09:08:31.267068] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.286 [2024-11-06 09:08:31.267285] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.286 [2024-11-06 09:08:31.267294] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.286 [2024-11-06 09:08:31.267302] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.286 [2024-11-06 09:08:31.270821] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.286 [2024-11-06 09:08:31.280168] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.286 [2024-11-06 09:08:31.280835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.286 [2024-11-06 09:08:31.280873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.286 [2024-11-06 09:08:31.280884] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.286 [2024-11-06 09:08:31.281121] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.286 [2024-11-06 09:08:31.281342] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.286 [2024-11-06 09:08:31.281351] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.286 [2024-11-06 09:08:31.281359] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.286 [2024-11-06 09:08:31.284898] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.286 [2024-11-06 09:08:31.294028] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.286 [2024-11-06 09:08:31.294702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.286 [2024-11-06 09:08:31.294740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.286 [2024-11-06 09:08:31.294761] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.286 [2024-11-06 09:08:31.295000] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.286 [2024-11-06 09:08:31.295221] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.286 [2024-11-06 09:08:31.295230] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.286 [2024-11-06 09:08:31.295238] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.286 [2024-11-06 09:08:31.298762] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.286 [2024-11-06 09:08:31.307896] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.286 [2024-11-06 09:08:31.308565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.286 [2024-11-06 09:08:31.308603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.286 [2024-11-06 09:08:31.308614] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.286 [2024-11-06 09:08:31.308866] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.286 [2024-11-06 09:08:31.309089] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.286 [2024-11-06 09:08:31.309098] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.286 [2024-11-06 09:08:31.309105] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.286 [2024-11-06 09:08:31.312626] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.286 [2024-11-06 09:08:31.321753] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.286 [2024-11-06 09:08:31.322384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.286 [2024-11-06 09:08:31.322421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.286 [2024-11-06 09:08:31.322432] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.286 [2024-11-06 09:08:31.322668] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.286 [2024-11-06 09:08:31.322899] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.286 [2024-11-06 09:08:31.322909] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.286 [2024-11-06 09:08:31.322917] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.286 [2024-11-06 09:08:31.326438] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.286 [2024-11-06 09:08:31.335565] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.286 [2024-11-06 09:08:31.336099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.286 [2024-11-06 09:08:31.336136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.286 [2024-11-06 09:08:31.336147] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.286 [2024-11-06 09:08:31.336384] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.286 [2024-11-06 09:08:31.336605] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.286 [2024-11-06 09:08:31.336614] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.286 [2024-11-06 09:08:31.336623] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.286 [2024-11-06 09:08:31.340153] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.286 [2024-11-06 09:08:31.349483] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.286 [2024-11-06 09:08:31.350085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.286 [2024-11-06 09:08:31.350123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.286 [2024-11-06 09:08:31.350134] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.286 [2024-11-06 09:08:31.350371] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.287 [2024-11-06 09:08:31.350593] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.287 [2024-11-06 09:08:31.350606] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.287 [2024-11-06 09:08:31.350613] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.287 [2024-11-06 09:08:31.354144] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.287 [2024-11-06 09:08:31.363265] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.287 [2024-11-06 09:08:31.363847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.287 [2024-11-06 09:08:31.363886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.287 [2024-11-06 09:08:31.363898] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.287 [2024-11-06 09:08:31.364136] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.287 [2024-11-06 09:08:31.364358] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.287 [2024-11-06 09:08:31.364367] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.287 [2024-11-06 09:08:31.364374] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.287 [2024-11-06 09:08:31.367904] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.287 [2024-11-06 09:08:31.377032] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.287 [2024-11-06 09:08:31.377702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.287 [2024-11-06 09:08:31.377740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.287 [2024-11-06 09:08:31.377761] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.287 [2024-11-06 09:08:31.377999] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.287 [2024-11-06 09:08:31.378221] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.287 [2024-11-06 09:08:31.378230] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.287 [2024-11-06 09:08:31.378238] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.287 [2024-11-06 09:08:31.381774] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.287 [2024-11-06 09:08:31.390913] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.287 [2024-11-06 09:08:31.391583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.287 [2024-11-06 09:08:31.391621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.287 [2024-11-06 09:08:31.391631] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.287 [2024-11-06 09:08:31.391877] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.287 [2024-11-06 09:08:31.392099] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.287 [2024-11-06 09:08:31.392109] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.287 [2024-11-06 09:08:31.392116] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.548 [2024-11-06 09:08:31.395644] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.548 [2024-11-06 09:08:31.404796] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.548 [2024-11-06 09:08:31.405470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.548 [2024-11-06 09:08:31.405508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.548 [2024-11-06 09:08:31.405518] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.548 [2024-11-06 09:08:31.405762] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.548 [2024-11-06 09:08:31.405985] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.549 [2024-11-06 09:08:31.405994] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.549 [2024-11-06 09:08:31.406003] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.549 [2024-11-06 09:08:31.409523] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.549 [2024-11-06 09:08:31.418683] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.549 [2024-11-06 09:08:31.419321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.549 [2024-11-06 09:08:31.419359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.549 [2024-11-06 09:08:31.419370] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.549 [2024-11-06 09:08:31.419607] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.549 [2024-11-06 09:08:31.419836] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.549 [2024-11-06 09:08:31.419845] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.549 [2024-11-06 09:08:31.419853] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.549 [2024-11-06 09:08:31.423374] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.549 [2024-11-06 09:08:31.432498] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.549 [2024-11-06 09:08:31.433094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.549 [2024-11-06 09:08:31.433113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.549 [2024-11-06 09:08:31.433121] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.549 [2024-11-06 09:08:31.433340] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.549 [2024-11-06 09:08:31.433557] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.549 [2024-11-06 09:08:31.433566] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.549 [2024-11-06 09:08:31.433573] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.549 [2024-11-06 09:08:31.437094] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.549 [2024-11-06 09:08:31.446420] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.549 [2024-11-06 09:08:31.447064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.549 [2024-11-06 09:08:31.447107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.549 [2024-11-06 09:08:31.447118] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.549 [2024-11-06 09:08:31.447355] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.549 [2024-11-06 09:08:31.447576] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.549 [2024-11-06 09:08:31.447585] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.549 [2024-11-06 09:08:31.447592] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.549 [2024-11-06 09:08:31.451122] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.549 [2024-11-06 09:08:31.460256] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.549 [2024-11-06 09:08:31.460810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.549 [2024-11-06 09:08:31.460848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.549 [2024-11-06 09:08:31.460860] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.549 [2024-11-06 09:08:31.461101] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.549 [2024-11-06 09:08:31.461322] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.549 [2024-11-06 09:08:31.461331] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.549 [2024-11-06 09:08:31.461339] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.549 [2024-11-06 09:08:31.464871] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.549 [2024-11-06 09:08:31.474205] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.549 [2024-11-06 09:08:31.474959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.549 [2024-11-06 09:08:31.474997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.549 [2024-11-06 09:08:31.475009] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.549 [2024-11-06 09:08:31.475245] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.549 [2024-11-06 09:08:31.475467] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.549 [2024-11-06 09:08:31.475476] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.549 [2024-11-06 09:08:31.475483] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.549 [2024-11-06 09:08:31.479015] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.549 [2024-11-06 09:08:31.488152] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.549 [2024-11-06 09:08:31.488827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.549 [2024-11-06 09:08:31.488865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.549 [2024-11-06 09:08:31.488876] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.549 [2024-11-06 09:08:31.489118] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.549 [2024-11-06 09:08:31.489340] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.549 [2024-11-06 09:08:31.489348] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.549 [2024-11-06 09:08:31.489356] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.549 [2024-11-06 09:08:31.492902] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.549 [2024-11-06 09:08:31.502056] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.549 [2024-11-06 09:08:31.502682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.549 [2024-11-06 09:08:31.502720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.549 [2024-11-06 09:08:31.502730] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.549 [2024-11-06 09:08:31.502976] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.549 [2024-11-06 09:08:31.503199] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.549 [2024-11-06 09:08:31.503207] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.549 [2024-11-06 09:08:31.503215] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.549 [2024-11-06 09:08:31.506738] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.549 [2024-11-06 09:08:31.515876] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.549 [2024-11-06 09:08:31.516520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.549 [2024-11-06 09:08:31.516558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.549 [2024-11-06 09:08:31.516569] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.549 [2024-11-06 09:08:31.516814] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.549 [2024-11-06 09:08:31.517036] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.549 [2024-11-06 09:08:31.517046] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.549 [2024-11-06 09:08:31.517054] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.549 [2024-11-06 09:08:31.520577] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.549 [2024-11-06 09:08:31.529703] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.549 [2024-11-06 09:08:31.530260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.549 [2024-11-06 09:08:31.530298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.549 [2024-11-06 09:08:31.530309] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.549 [2024-11-06 09:08:31.530546] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.549 [2024-11-06 09:08:31.530775] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.549 [2024-11-06 09:08:31.530789] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.549 [2024-11-06 09:08:31.530798] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.549 [2024-11-06 09:08:31.534319] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.549 [2024-11-06 09:08:31.543653] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.549 [2024-11-06 09:08:31.544312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.549 [2024-11-06 09:08:31.544350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.549 [2024-11-06 09:08:31.544363] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.550 [2024-11-06 09:08:31.544601] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.550 [2024-11-06 09:08:31.544832] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.550 [2024-11-06 09:08:31.544842] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.550 [2024-11-06 09:08:31.544850] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.550 [2024-11-06 09:08:31.548372] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.550 [2024-11-06 09:08:31.557498] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.550 [2024-11-06 09:08:31.558069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.550 [2024-11-06 09:08:31.558108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.550 [2024-11-06 09:08:31.558119] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.550 [2024-11-06 09:08:31.558356] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.550 [2024-11-06 09:08:31.558577] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.550 [2024-11-06 09:08:31.558586] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.550 [2024-11-06 09:08:31.558593] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.550 [2024-11-06 09:08:31.562122] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.550 [2024-11-06 09:08:31.571451] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.550 [2024-11-06 09:08:31.572133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.550 [2024-11-06 09:08:31.572171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.550 [2024-11-06 09:08:31.572182] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.550 [2024-11-06 09:08:31.572418] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.550 [2024-11-06 09:08:31.572640] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.550 [2024-11-06 09:08:31.572649] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.550 [2024-11-06 09:08:31.572657] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.550 [2024-11-06 09:08:31.576187] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.550 [2024-11-06 09:08:31.585334] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.550 [2024-11-06 09:08:31.586037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.550 [2024-11-06 09:08:31.586075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.550 [2024-11-06 09:08:31.586086] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.550 [2024-11-06 09:08:31.586323] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.550 [2024-11-06 09:08:31.586545] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.550 [2024-11-06 09:08:31.586554] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.550 [2024-11-06 09:08:31.586562] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.550 [2024-11-06 09:08:31.590090] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.550 [2024-11-06 09:08:31.599221] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.550 [2024-11-06 09:08:31.599848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.550 [2024-11-06 09:08:31.599885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.550 [2024-11-06 09:08:31.599897] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.550 [2024-11-06 09:08:31.600138] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.550 [2024-11-06 09:08:31.600359] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.550 [2024-11-06 09:08:31.600368] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.550 [2024-11-06 09:08:31.600376] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.550 [2024-11-06 09:08:31.603917] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.550 [2024-11-06 09:08:31.613042] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.550 [2024-11-06 09:08:31.613707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.550 [2024-11-06 09:08:31.613753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.550 [2024-11-06 09:08:31.613765] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.550 [2024-11-06 09:08:31.614002] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.550 [2024-11-06 09:08:31.614224] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.550 [2024-11-06 09:08:31.614233] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.550 [2024-11-06 09:08:31.614240] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.550 [2024-11-06 09:08:31.617764] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.550 [2024-11-06 09:08:31.626885] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.550 [2024-11-06 09:08:31.627550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.550 [2024-11-06 09:08:31.627592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.550 [2024-11-06 09:08:31.627603] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.550 [2024-11-06 09:08:31.627850] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.550 [2024-11-06 09:08:31.628073] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.550 [2024-11-06 09:08:31.628081] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.550 [2024-11-06 09:08:31.628089] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.550 [2024-11-06 09:08:31.631611] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.550 [2024-11-06 09:08:31.640734] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.550 [2024-11-06 09:08:31.641372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.550 [2024-11-06 09:08:31.641411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.550 [2024-11-06 09:08:31.641422] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.550 [2024-11-06 09:08:31.641659] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.550 [2024-11-06 09:08:31.641889] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.550 [2024-11-06 09:08:31.641899] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.550 [2024-11-06 09:08:31.641907] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.550 [2024-11-06 09:08:31.645428] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.550 [2024-11-06 09:08:31.654551] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.550 [2024-11-06 09:08:31.655224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.550 [2024-11-06 09:08:31.655262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.550 [2024-11-06 09:08:31.655272] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.550 [2024-11-06 09:08:31.655509] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.550 [2024-11-06 09:08:31.655730] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.550 [2024-11-06 09:08:31.655740] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.550 [2024-11-06 09:08:31.655757] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.812 [2024-11-06 09:08:31.659280] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.812 [2024-11-06 09:08:31.668408] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.812 [2024-11-06 09:08:31.669042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.812 [2024-11-06 09:08:31.669081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.812 [2024-11-06 09:08:31.669091] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.812 [2024-11-06 09:08:31.669333] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.812 [2024-11-06 09:08:31.669555] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.812 [2024-11-06 09:08:31.669564] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.812 [2024-11-06 09:08:31.669571] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.812 [2024-11-06 09:08:31.673102] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.812 [2024-11-06 09:08:31.682237] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.812 [2024-11-06 09:08:31.682849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.812 [2024-11-06 09:08:31.682887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.812 [2024-11-06 09:08:31.682900] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.812 [2024-11-06 09:08:31.683138] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.812 [2024-11-06 09:08:31.683360] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.812 [2024-11-06 09:08:31.683369] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.812 [2024-11-06 09:08:31.683376] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.812 [2024-11-06 09:08:31.686906] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.812 [2024-11-06 09:08:31.696039] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.812 [2024-11-06 09:08:31.696710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.812 [2024-11-06 09:08:31.696755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.812 [2024-11-06 09:08:31.696768] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.812 [2024-11-06 09:08:31.697006] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.812 [2024-11-06 09:08:31.697228] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.812 [2024-11-06 09:08:31.697237] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.812 [2024-11-06 09:08:31.697245] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.812 [2024-11-06 09:08:31.700791] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.812 [2024-11-06 09:08:31.709925] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.812 [2024-11-06 09:08:31.710616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.812 [2024-11-06 09:08:31.710654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.812 [2024-11-06 09:08:31.710664] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.812 [2024-11-06 09:08:31.710912] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.711135] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.711143] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.711156] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.813 [2024-11-06 09:08:31.714677] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.813 [2024-11-06 09:08:31.723801] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.813 [2024-11-06 09:08:31.724437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.813 [2024-11-06 09:08:31.724475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.813 [2024-11-06 09:08:31.724486] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.813 [2024-11-06 09:08:31.724723] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.724953] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.724963] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.724971] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.813 [2024-11-06 09:08:31.728495] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.813 [2024-11-06 09:08:31.737630] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.813 [2024-11-06 09:08:31.738264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.813 [2024-11-06 09:08:31.738302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.813 [2024-11-06 09:08:31.738313] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.813 [2024-11-06 09:08:31.738550] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.738780] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.738790] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.738798] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.813 [2024-11-06 09:08:31.742318] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.813 [2024-11-06 09:08:31.751443] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.813 [2024-11-06 09:08:31.752095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.813 [2024-11-06 09:08:31.752134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.813 [2024-11-06 09:08:31.752145] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.813 [2024-11-06 09:08:31.752382] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.752603] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.752612] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.752620] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.813 [2024-11-06 09:08:31.756147] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.813 [2024-11-06 09:08:31.765283] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.813 [2024-11-06 09:08:31.765893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.813 [2024-11-06 09:08:31.765913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.813 [2024-11-06 09:08:31.765922] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.813 [2024-11-06 09:08:31.766140] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.766359] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.766368] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.766376] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.813 [2024-11-06 09:08:31.769896] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.813 [2024-11-06 09:08:31.779234] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.813 [2024-11-06 09:08:31.779947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.813 [2024-11-06 09:08:31.779986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.813 [2024-11-06 09:08:31.779996] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.813 [2024-11-06 09:08:31.780234] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.780466] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.780477] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.780485] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.813 [2024-11-06 09:08:31.784015] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.813 [2024-11-06 09:08:31.793228] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.813 [2024-11-06 09:08:31.793846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.813 [2024-11-06 09:08:31.793884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.813 [2024-11-06 09:08:31.793896] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.813 [2024-11-06 09:08:31.794134] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.794356] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.794365] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.794373] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.813 [2024-11-06 09:08:31.797904] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.813 [2024-11-06 09:08:31.807040] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.813 [2024-11-06 09:08:31.807713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.813 [2024-11-06 09:08:31.807763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.813 [2024-11-06 09:08:31.807775] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.813 [2024-11-06 09:08:31.808012] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.808233] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.808242] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.808250] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.813 [2024-11-06 09:08:31.811777] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.813 [2024-11-06 09:08:31.820902] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.813 [2024-11-06 09:08:31.821540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.813 [2024-11-06 09:08:31.821578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.813 [2024-11-06 09:08:31.821589] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.813 [2024-11-06 09:08:31.821834] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.822057] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.822066] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.822074] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.813 [2024-11-06 09:08:31.825595] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.813 [2024-11-06 09:08:31.834723] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.813 [2024-11-06 09:08:31.835397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.813 [2024-11-06 09:08:31.835435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.813 [2024-11-06 09:08:31.835446] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.813 [2024-11-06 09:08:31.835683] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.835913] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.835923] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.835930] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.813 [2024-11-06 09:08:31.839448] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.813 [2024-11-06 09:08:31.848576] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.813 [2024-11-06 09:08:31.849139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.813 [2024-11-06 09:08:31.849177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.813 [2024-11-06 09:08:31.849188] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.813 [2024-11-06 09:08:31.849424] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.813 [2024-11-06 09:08:31.849650] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.813 [2024-11-06 09:08:31.849660] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.813 [2024-11-06 09:08:31.849668] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.814 [2024-11-06 09:08:31.853199] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.814 [2024-11-06 09:08:31.862529] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.814 [2024-11-06 09:08:31.863173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.814 [2024-11-06 09:08:31.863211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.814 [2024-11-06 09:08:31.863222] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.814 [2024-11-06 09:08:31.863459] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.814 [2024-11-06 09:08:31.863680] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.814 [2024-11-06 09:08:31.863688] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.814 [2024-11-06 09:08:31.863696] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.814 [2024-11-06 09:08:31.867226] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.814 [2024-11-06 09:08:31.876362] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.814 [2024-11-06 09:08:31.877057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.814 [2024-11-06 09:08:31.877095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.814 [2024-11-06 09:08:31.877106] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.814 [2024-11-06 09:08:31.877343] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.814 [2024-11-06 09:08:31.877564] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.814 [2024-11-06 09:08:31.877573] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.814 [2024-11-06 09:08:31.877581] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.814 [2024-11-06 09:08:31.881122] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.814 [2024-11-06 09:08:31.890258] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.814 [2024-11-06 09:08:31.890849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.814 [2024-11-06 09:08:31.890887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.814 [2024-11-06 09:08:31.890900] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.814 [2024-11-06 09:08:31.891141] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.814 [2024-11-06 09:08:31.891363] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.814 [2024-11-06 09:08:31.891371] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.814 [2024-11-06 09:08:31.891383] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.814 [2024-11-06 09:08:31.894914] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.814 [2024-11-06 09:08:31.904047] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.814 [2024-11-06 09:08:31.904720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.814 [2024-11-06 09:08:31.904765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.814 [2024-11-06 09:08:31.904777] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.814 [2024-11-06 09:08:31.905014] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.814 [2024-11-06 09:08:31.905235] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.814 [2024-11-06 09:08:31.905244] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.814 [2024-11-06 09:08:31.905252] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.814 [2024-11-06 09:08:31.908795] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:41.814 [2024-11-06 09:08:31.917928] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:41.814 [2024-11-06 09:08:31.918512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:41.814 [2024-11-06 09:08:31.918531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:41.814 [2024-11-06 09:08:31.918539] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:41.814 [2024-11-06 09:08:31.918763] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:41.814 [2024-11-06 09:08:31.918981] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:41.814 [2024-11-06 09:08:31.918989] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:41.814 [2024-11-06 09:08:31.918997] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:41.814 [2024-11-06 09:08:31.922510] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.076 [2024-11-06 09:08:31.931839] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.076 [2024-11-06 09:08:31.932409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.076 [2024-11-06 09:08:31.932425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.076 [2024-11-06 09:08:31.932433] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.076 [2024-11-06 09:08:31.932650] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.076 [2024-11-06 09:08:31.932872] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.076 [2024-11-06 09:08:31.932881] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.076 [2024-11-06 09:08:31.932888] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.076 [2024-11-06 09:08:31.936401] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.076 [2024-11-06 09:08:31.945735] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.076 [2024-11-06 09:08:31.946263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.076 [2024-11-06 09:08:31.946280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.076 [2024-11-06 09:08:31.946287] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.076 [2024-11-06 09:08:31.946504] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.076 [2024-11-06 09:08:31.946721] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.076 [2024-11-06 09:08:31.946730] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.076 [2024-11-06 09:08:31.946737] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.076 [2024-11-06 09:08:31.950255] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.076 [2024-11-06 09:08:31.959580] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.076 [2024-11-06 09:08:31.960137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.076 [2024-11-06 09:08:31.960153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.076 [2024-11-06 09:08:31.960161] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.076 [2024-11-06 09:08:31.960377] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.076 [2024-11-06 09:08:31.960594] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.076 [2024-11-06 09:08:31.960603] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.076 [2024-11-06 09:08:31.960610] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.076 [2024-11-06 09:08:31.964130] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.076 [2024-11-06 09:08:31.973460] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.076 [2024-11-06 09:08:31.974132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.076 [2024-11-06 09:08:31.974171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.076 [2024-11-06 09:08:31.974182] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.076 [2024-11-06 09:08:31.974419] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.076 [2024-11-06 09:08:31.974640] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.076 [2024-11-06 09:08:31.974650] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.076 [2024-11-06 09:08:31.974658] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.076 [2024-11-06 09:08:31.978185] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.076 [2024-11-06 09:08:31.987324] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.076 [2024-11-06 09:08:31.988043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.076 [2024-11-06 09:08:31.988081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.076 [2024-11-06 09:08:31.988096] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.076 [2024-11-06 09:08:31.988333] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.076 [2024-11-06 09:08:31.988555] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.076 [2024-11-06 09:08:31.988564] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.076 [2024-11-06 09:08:31.988572] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.076 [2024-11-06 09:08:31.992100] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.076 [2024-11-06 09:08:32.001238] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.076 [2024-11-06 09:08:32.001852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.076 [2024-11-06 09:08:32.001890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.076 [2024-11-06 09:08:32.001903] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.076 [2024-11-06 09:08:32.002141] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.076 [2024-11-06 09:08:32.002363] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.076 [2024-11-06 09:08:32.002373] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.076 [2024-11-06 09:08:32.002381] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.076 [2024-11-06 09:08:32.005911] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.076 [2024-11-06 09:08:32.015035] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.076 [2024-11-06 09:08:32.015619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.076 [2024-11-06 09:08:32.015638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.076 [2024-11-06 09:08:32.015646] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.076 [2024-11-06 09:08:32.015869] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.076 [2024-11-06 09:08:32.016088] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.076 [2024-11-06 09:08:32.016097] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.076 [2024-11-06 09:08:32.016105] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.076 [2024-11-06 09:08:32.019621] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.076 [2024-11-06 09:08:32.028969] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.076 [2024-11-06 09:08:32.029575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.076 [2024-11-06 09:08:32.029614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.076 [2024-11-06 09:08:32.029626] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.076 [2024-11-06 09:08:32.029875] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.076 [2024-11-06 09:08:32.030103] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.076 [2024-11-06 09:08:32.030113] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.076 [2024-11-06 09:08:32.030121] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.076 [2024-11-06 09:08:32.033655] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.076 [2024-11-06 09:08:32.042785] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.076 [2024-11-06 09:08:32.043465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.076 [2024-11-06 09:08:32.043503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.076 [2024-11-06 09:08:32.043514] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.076 [2024-11-06 09:08:32.043759] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.076 [2024-11-06 09:08:32.043982] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.076 [2024-11-06 09:08:32.043991] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.076 [2024-11-06 09:08:32.043999] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.076 [2024-11-06 09:08:32.047520] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.076 [2024-11-06 09:08:32.056653] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.077 [2024-11-06 09:08:32.057247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.077 [2024-11-06 09:08:32.057267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.077 [2024-11-06 09:08:32.057275] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.077 [2024-11-06 09:08:32.057493] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.077 [2024-11-06 09:08:32.057710] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.077 [2024-11-06 09:08:32.057718] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.077 [2024-11-06 09:08:32.057725] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.077 [2024-11-06 09:08:32.061250] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.077 [2024-11-06 09:08:32.070587] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.077 [2024-11-06 09:08:32.071121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.077 [2024-11-06 09:08:32.071137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.077 [2024-11-06 09:08:32.071144] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.077 [2024-11-06 09:08:32.071362] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.077 [2024-11-06 09:08:32.071579] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.077 [2024-11-06 09:08:32.071588] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.077 [2024-11-06 09:08:32.071600] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.077 [2024-11-06 09:08:32.075120] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.077 [2024-11-06 09:08:32.084466] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.077 [2024-11-06 09:08:32.085034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.077 [2024-11-06 09:08:32.085051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.077 [2024-11-06 09:08:32.085059] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.077 [2024-11-06 09:08:32.085276] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.077 [2024-11-06 09:08:32.085493] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.077 [2024-11-06 09:08:32.085502] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.077 [2024-11-06 09:08:32.085509] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.077 [2024-11-06 09:08:32.089033] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.077 [2024-11-06 09:08:32.098373] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.077 [2024-11-06 09:08:32.098836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.077 [2024-11-06 09:08:32.098852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.077 [2024-11-06 09:08:32.098859] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.077 [2024-11-06 09:08:32.099076] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.077 [2024-11-06 09:08:32.099293] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.077 [2024-11-06 09:08:32.099302] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.077 [2024-11-06 09:08:32.099309] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.077 [2024-11-06 09:08:32.102845] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.077 [2024-11-06 09:08:32.112191] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.077 [2024-11-06 09:08:32.112597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.077 [2024-11-06 09:08:32.112615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.077 [2024-11-06 09:08:32.112623] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.077 [2024-11-06 09:08:32.112849] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.077 [2024-11-06 09:08:32.113069] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.077 [2024-11-06 09:08:32.113077] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.077 [2024-11-06 09:08:32.113084] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.077 [2024-11-06 09:08:32.116697] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.077 [2024-11-06 09:08:32.126057] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.077 [2024-11-06 09:08:32.126724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.077 [2024-11-06 09:08:32.126768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.077 [2024-11-06 09:08:32.126779] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.077 [2024-11-06 09:08:32.127016] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.077 [2024-11-06 09:08:32.127237] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.077 [2024-11-06 09:08:32.127246] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.077 [2024-11-06 09:08:32.127254] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.077 [2024-11-06 09:08:32.130778] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.077 [2024-11-06 09:08:32.139906] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.077 [2024-11-06 09:08:32.140462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.077 [2024-11-06 09:08:32.140481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.077 [2024-11-06 09:08:32.140489] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.077 [2024-11-06 09:08:32.140707] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.077 [2024-11-06 09:08:32.140931] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.077 [2024-11-06 09:08:32.140941] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.077 [2024-11-06 09:08:32.140948] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.077 [2024-11-06 09:08:32.144464] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.077 [2024-11-06 09:08:32.153794] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.077 [2024-11-06 09:08:32.154450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.077 [2024-11-06 09:08:32.154488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.077 [2024-11-06 09:08:32.154498] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.077 [2024-11-06 09:08:32.154736] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.077 [2024-11-06 09:08:32.154967] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.077 [2024-11-06 09:08:32.154977] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.077 [2024-11-06 09:08:32.154984] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.077 [2024-11-06 09:08:32.158507] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.077 [2024-11-06 09:08:32.167637] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.077 [2024-11-06 09:08:32.167951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.077 [2024-11-06 09:08:32.167974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.077 [2024-11-06 09:08:32.167986] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.077 [2024-11-06 09:08:32.168207] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.077 [2024-11-06 09:08:32.168425] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.077 [2024-11-06 09:08:32.168433] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.077 [2024-11-06 09:08:32.168440] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.077 [2024-11-06 09:08:32.171964] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.077 7122.75 IOPS, 27.82 MiB/s [2024-11-06T08:08:32.190Z] [2024-11-06 09:08:32.183193] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.077 [2024-11-06 09:08:32.183724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.077 [2024-11-06 09:08:32.183740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.077 [2024-11-06 09:08:32.183753] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.077 [2024-11-06 09:08:32.183971] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.077 [2024-11-06 09:08:32.184188] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.077 [2024-11-06 09:08:32.184196] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.077 [2024-11-06 09:08:32.184203] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.339 [2024-11-06 09:08:32.187722] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.339 [2024-11-06 09:08:32.197053] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.339 [2024-11-06 09:08:32.197654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.339 [2024-11-06 09:08:32.197692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.339 [2024-11-06 09:08:32.197705] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.339 [2024-11-06 09:08:32.197954] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.339 [2024-11-06 09:08:32.198177] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.339 [2024-11-06 09:08:32.198187] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.339 [2024-11-06 09:08:32.198195] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.339 [2024-11-06 09:08:32.201724] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.339 [2024-11-06 09:08:32.210868] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.339 [2024-11-06 09:08:32.211271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.339 [2024-11-06 09:08:32.211291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.339 [2024-11-06 09:08:32.211299] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.339 [2024-11-06 09:08:32.211518] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.339 [2024-11-06 09:08:32.211742] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.339 [2024-11-06 09:08:32.211758] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.339 [2024-11-06 09:08:32.211765] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.339 [2024-11-06 09:08:32.215283] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.339 [2024-11-06 09:08:32.224830] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.339 [2024-11-06 09:08:32.225395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.339 [2024-11-06 09:08:32.225411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.340 [2024-11-06 09:08:32.225418] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.340 [2024-11-06 09:08:32.225635] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.340 [2024-11-06 09:08:32.225859] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.340 [2024-11-06 09:08:32.225868] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.340 [2024-11-06 09:08:32.225875] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.340 [2024-11-06 09:08:32.229395] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.340 [2024-11-06 09:08:32.238735] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.340 [2024-11-06 09:08:32.239352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.340 [2024-11-06 09:08:32.239390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.340 [2024-11-06 09:08:32.239401] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.340 [2024-11-06 09:08:32.239638] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.340 [2024-11-06 09:08:32.239869] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.340 [2024-11-06 09:08:32.239878] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.340 [2024-11-06 09:08:32.239886] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.340 [2024-11-06 09:08:32.243412] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.340 [2024-11-06 09:08:32.252552] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.340 [2024-11-06 09:08:32.253101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.340 [2024-11-06 09:08:32.253121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.340 [2024-11-06 09:08:32.253129] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.340 [2024-11-06 09:08:32.253346] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.340 [2024-11-06 09:08:32.253564] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.340 [2024-11-06 09:08:32.253572] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.340 [2024-11-06 09:08:32.253583] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.340 [2024-11-06 09:08:32.257112] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.340 [2024-11-06 09:08:32.266454] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.340 [2024-11-06 09:08:32.266978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.340 [2024-11-06 09:08:32.266995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.340 [2024-11-06 09:08:32.267003] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.340 [2024-11-06 09:08:32.267220] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.340 [2024-11-06 09:08:32.267437] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.340 [2024-11-06 09:08:32.267445] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.340 [2024-11-06 09:08:32.267452] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.340 [2024-11-06 09:08:32.270976] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.340 [2024-11-06 09:08:32.280316] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.340 [2024-11-06 09:08:32.280822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.340 [2024-11-06 09:08:32.280838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.340 [2024-11-06 09:08:32.280846] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.340 [2024-11-06 09:08:32.281063] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.340 [2024-11-06 09:08:32.281280] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.340 [2024-11-06 09:08:32.281289] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.340 [2024-11-06 09:08:32.281297] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.340 [2024-11-06 09:08:32.284829] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.340 [2024-11-06 09:08:32.294176] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.340 [2024-11-06 09:08:32.294715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.340 [2024-11-06 09:08:32.294731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.340 [2024-11-06 09:08:32.294739] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.340 [2024-11-06 09:08:32.294962] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.340 [2024-11-06 09:08:32.295179] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.340 [2024-11-06 09:08:32.295187] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.340 [2024-11-06 09:08:32.295194] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.340 [2024-11-06 09:08:32.298711] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.340 [2024-11-06 09:08:32.308075] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.340 [2024-11-06 09:08:32.308595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.340 [2024-11-06 09:08:32.308610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.340 [2024-11-06 09:08:32.308618] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.340 [2024-11-06 09:08:32.308841] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.340 [2024-11-06 09:08:32.309059] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.340 [2024-11-06 09:08:32.309067] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.340 [2024-11-06 09:08:32.309075] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.340 [2024-11-06 09:08:32.312594] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.340 [2024-11-06 09:08:32.321940] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.340 [2024-11-06 09:08:32.322462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.340 [2024-11-06 09:08:32.322478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.340 [2024-11-06 09:08:32.322486] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.340 [2024-11-06 09:08:32.322702] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.340 [2024-11-06 09:08:32.322935] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.340 [2024-11-06 09:08:32.322947] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.340 [2024-11-06 09:08:32.322954] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.340 [2024-11-06 09:08:32.326475] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.340 [2024-11-06 09:08:32.335821] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.340 [2024-11-06 09:08:32.336211] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.340 [2024-11-06 09:08:32.336229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.340 [2024-11-06 09:08:32.336237] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.340 [2024-11-06 09:08:32.336455] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.340 [2024-11-06 09:08:32.336671] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.340 [2024-11-06 09:08:32.336679] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.340 [2024-11-06 09:08:32.336686] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.340 [2024-11-06 09:08:32.340213] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.340 [2024-11-06 09:08:32.349767] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.340 [2024-11-06 09:08:32.350336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.340 [2024-11-06 09:08:32.350351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.340 [2024-11-06 09:08:32.350363] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.340 [2024-11-06 09:08:32.350580] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.340 [2024-11-06 09:08:32.350804] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.340 [2024-11-06 09:08:32.350814] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.340 [2024-11-06 09:08:32.350821] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.340 [2024-11-06 09:08:32.354337] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.340 [2024-11-06 09:08:32.363678] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.340 [2024-11-06 09:08:32.364212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.340 [2024-11-06 09:08:32.364227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.341 [2024-11-06 09:08:32.364235] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.341 [2024-11-06 09:08:32.364452] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.341 [2024-11-06 09:08:32.364668] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.341 [2024-11-06 09:08:32.364676] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.341 [2024-11-06 09:08:32.364683] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.341 [2024-11-06 09:08:32.368206] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.341 [2024-11-06 09:08:32.377544] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.341 [2024-11-06 09:08:32.378100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.341 [2024-11-06 09:08:32.378115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.341 [2024-11-06 09:08:32.378123] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.341 [2024-11-06 09:08:32.378340] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.341 [2024-11-06 09:08:32.378556] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.341 [2024-11-06 09:08:32.378565] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.341 [2024-11-06 09:08:32.378572] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.341 [2024-11-06 09:08:32.382095] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.341 [2024-11-06 09:08:32.391442] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.341 [2024-11-06 09:08:32.391864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.341 [2024-11-06 09:08:32.391881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.341 [2024-11-06 09:08:32.391889] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.341 [2024-11-06 09:08:32.392107] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.341 [2024-11-06 09:08:32.392328] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.341 [2024-11-06 09:08:32.392337] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.341 [2024-11-06 09:08:32.392344] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.341 [2024-11-06 09:08:32.395867] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.341 [2024-11-06 09:08:32.405215] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.341 [2024-11-06 09:08:32.405742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.341 [2024-11-06 09:08:32.405763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.341 [2024-11-06 09:08:32.405771] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.341 [2024-11-06 09:08:32.405988] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.341 [2024-11-06 09:08:32.406205] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.341 [2024-11-06 09:08:32.406213] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.341 [2024-11-06 09:08:32.406220] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.341 [2024-11-06 09:08:32.409738] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.341 [2024-11-06 09:08:32.419082] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.341 [2024-11-06 09:08:32.419643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.341 [2024-11-06 09:08:32.419659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.341 [2024-11-06 09:08:32.419666] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.341 [2024-11-06 09:08:32.419889] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.341 [2024-11-06 09:08:32.420107] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.341 [2024-11-06 09:08:32.420115] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.341 [2024-11-06 09:08:32.420122] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.341 [2024-11-06 09:08:32.423639] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.341 [2024-11-06 09:08:32.432980] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.341 [2024-11-06 09:08:32.433504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.341 [2024-11-06 09:08:32.433519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.341 [2024-11-06 09:08:32.433527] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.341 [2024-11-06 09:08:32.433744] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.341 [2024-11-06 09:08:32.433968] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.341 [2024-11-06 09:08:32.433976] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.341 [2024-11-06 09:08:32.433990] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.341 [2024-11-06 09:08:32.437510] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.341 [2024-11-06 09:08:32.446851] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.341 [2024-11-06 09:08:32.447400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.341 [2024-11-06 09:08:32.447415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.341 [2024-11-06 09:08:32.447423] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.341 [2024-11-06 09:08:32.447639] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.341 [2024-11-06 09:08:32.447862] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.341 [2024-11-06 09:08:32.447871] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.341 [2024-11-06 09:08:32.447878] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.603 [2024-11-06 09:08:32.451395] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.603 [2024-11-06 09:08:32.460737] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.603 [2024-11-06 09:08:32.461273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.603 [2024-11-06 09:08:32.461289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.603 [2024-11-06 09:08:32.461296] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.603 [2024-11-06 09:08:32.461513] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.603 [2024-11-06 09:08:32.461730] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.603 [2024-11-06 09:08:32.461739] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.603 [2024-11-06 09:08:32.461752] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.603 [2024-11-06 09:08:32.465268] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.603 [2024-11-06 09:08:32.474617] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.603 [2024-11-06 09:08:32.475144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.603 [2024-11-06 09:08:32.475159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.603 [2024-11-06 09:08:32.475167] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.603 [2024-11-06 09:08:32.475384] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.603 [2024-11-06 09:08:32.475601] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.603 [2024-11-06 09:08:32.475610] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.603 [2024-11-06 09:08:32.475617] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.603 [2024-11-06 09:08:32.479144] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.603 [2024-11-06 09:08:32.488510] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.603 [2024-11-06 09:08:32.489035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.603 [2024-11-06 09:08:32.489051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.603 [2024-11-06 09:08:32.489058] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.603 [2024-11-06 09:08:32.489276] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.603 [2024-11-06 09:08:32.489493] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.603 [2024-11-06 09:08:32.489502] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.603 [2024-11-06 09:08:32.489509] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.603 [2024-11-06 09:08:32.493036] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.603 [2024-11-06 09:08:32.502396] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.603 [2024-11-06 09:08:32.502921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.603 [2024-11-06 09:08:32.502938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.603 [2024-11-06 09:08:32.502945] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.603 [2024-11-06 09:08:32.503162] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.603 [2024-11-06 09:08:32.503379] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.603 [2024-11-06 09:08:32.503388] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.603 [2024-11-06 09:08:32.503395] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.603 [2024-11-06 09:08:32.506923] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.603 [2024-11-06 09:08:32.516271] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.603 [2024-11-06 09:08:32.516971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.603 [2024-11-06 09:08:32.517009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.603 [2024-11-06 09:08:32.517020] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.603 [2024-11-06 09:08:32.517257] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.603 [2024-11-06 09:08:32.517479] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.603 [2024-11-06 09:08:32.517488] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.603 [2024-11-06 09:08:32.517496] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.603 [2024-11-06 09:08:32.521031] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.603 [2024-11-06 09:08:32.530175] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.603 [2024-11-06 09:08:32.530728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.603 [2024-11-06 09:08:32.530780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.603 [2024-11-06 09:08:32.530802] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.604 [2024-11-06 09:08:32.531041] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.604 [2024-11-06 09:08:32.531262] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.604 [2024-11-06 09:08:32.531277] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.604 [2024-11-06 09:08:32.531285] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.604 [2024-11-06 09:08:32.534818] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.604 [2024-11-06 09:08:32.543958] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.604 [2024-11-06 09:08:32.544544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.604 [2024-11-06 09:08:32.544563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.604 [2024-11-06 09:08:32.544571] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.604 [2024-11-06 09:08:32.544796] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.604 [2024-11-06 09:08:32.545015] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.604 [2024-11-06 09:08:32.545023] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.604 [2024-11-06 09:08:32.545030] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.604 [2024-11-06 09:08:32.548551] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.604 [2024-11-06 09:08:32.557899] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.604 [2024-11-06 09:08:32.558472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.604 [2024-11-06 09:08:32.558489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.604 [2024-11-06 09:08:32.558496] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.604 [2024-11-06 09:08:32.558713] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.604 [2024-11-06 09:08:32.558936] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.604 [2024-11-06 09:08:32.558945] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.604 [2024-11-06 09:08:32.558952] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.604 [2024-11-06 09:08:32.562469] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.604 [2024-11-06 09:08:32.571815] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.604 [2024-11-06 09:08:32.572338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.604 [2024-11-06 09:08:32.572353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.604 [2024-11-06 09:08:32.572361] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.604 [2024-11-06 09:08:32.572577] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.604 [2024-11-06 09:08:32.572805] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.604 [2024-11-06 09:08:32.572815] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.604 [2024-11-06 09:08:32.572822] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.604 [2024-11-06 09:08:32.576339] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.604 [2024-11-06 09:08:32.585690] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.604 [2024-11-06 09:08:32.586357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.604 [2024-11-06 09:08:32.586396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.604 [2024-11-06 09:08:32.586407] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.604 [2024-11-06 09:08:32.586644] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.604 [2024-11-06 09:08:32.586877] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.604 [2024-11-06 09:08:32.586887] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.604 [2024-11-06 09:08:32.586894] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.604 [2024-11-06 09:08:32.590418] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.604 [2024-11-06 09:08:32.599559] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.604 [2024-11-06 09:08:32.600147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.604 [2024-11-06 09:08:32.600167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.604 [2024-11-06 09:08:32.600175] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.604 [2024-11-06 09:08:32.600393] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.604 [2024-11-06 09:08:32.600610] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.604 [2024-11-06 09:08:32.600619] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.604 [2024-11-06 09:08:32.600626] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.604 [2024-11-06 09:08:32.604161] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.604 [2024-11-06 09:08:32.613502] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.604 [2024-11-06 09:08:32.614131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.604 [2024-11-06 09:08:32.614169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.604 [2024-11-06 09:08:32.614179] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.604 [2024-11-06 09:08:32.614417] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.604 [2024-11-06 09:08:32.614639] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.604 [2024-11-06 09:08:32.614648] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.604 [2024-11-06 09:08:32.614660] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.604 [2024-11-06 09:08:32.618195] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.604 [2024-11-06 09:08:32.627343] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.604 [2024-11-06 09:08:32.627983] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.604 [2024-11-06 09:08:32.628021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.604 [2024-11-06 09:08:32.628032] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.604 [2024-11-06 09:08:32.628269] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.604 [2024-11-06 09:08:32.628490] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.604 [2024-11-06 09:08:32.628499] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.604 [2024-11-06 09:08:32.628507] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.604 [2024-11-06 09:08:32.632036] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.604 [2024-11-06 09:08:32.641179] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.604 [2024-11-06 09:08:32.641866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.604 [2024-11-06 09:08:32.641905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.604 [2024-11-06 09:08:32.641916] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.604 [2024-11-06 09:08:32.642153] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.604 [2024-11-06 09:08:32.642374] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.604 [2024-11-06 09:08:32.642383] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.604 [2024-11-06 09:08:32.642391] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.604 [2024-11-06 09:08:32.645921] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.604 [2024-11-06 09:08:32.655052] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.604 [2024-11-06 09:08:32.655616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.604 [2024-11-06 09:08:32.655652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.604 [2024-11-06 09:08:32.655664] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.604 [2024-11-06 09:08:32.655909] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.604 [2024-11-06 09:08:32.656131] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.604 [2024-11-06 09:08:32.656139] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.604 [2024-11-06 09:08:32.656147] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.604 [2024-11-06 09:08:32.659668] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.604 [2024-11-06 09:08:32.669014] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.604 [2024-11-06 09:08:32.669610] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.604 [2024-11-06 09:08:32.669628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.605 [2024-11-06 09:08:32.669636] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.605 [2024-11-06 09:08:32.669861] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.605 [2024-11-06 09:08:32.670079] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.605 [2024-11-06 09:08:32.670087] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.605 [2024-11-06 09:08:32.670095] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.605 [2024-11-06 09:08:32.673615] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.605 [2024-11-06 09:08:32.682954] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.605 [2024-11-06 09:08:32.683501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.605 [2024-11-06 09:08:32.683537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.605 [2024-11-06 09:08:32.683549] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.605 [2024-11-06 09:08:32.683794] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.605 [2024-11-06 09:08:32.684016] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.605 [2024-11-06 09:08:32.684025] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.605 [2024-11-06 09:08:32.684033] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.605 [2024-11-06 09:08:32.687569] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.605 [2024-11-06 09:08:32.696913] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.605 [2024-11-06 09:08:32.697551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.605 [2024-11-06 09:08:32.697589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.605 [2024-11-06 09:08:32.697600] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.605 [2024-11-06 09:08:32.697847] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.605 [2024-11-06 09:08:32.698069] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.605 [2024-11-06 09:08:32.698078] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.605 [2024-11-06 09:08:32.698086] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.605 [2024-11-06 09:08:32.701617] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.605 [2024-11-06 09:08:32.710761] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.605 [2024-11-06 09:08:32.711402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.605 [2024-11-06 09:08:32.711440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.605 [2024-11-06 09:08:32.711455] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.605 [2024-11-06 09:08:32.711692] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.605 [2024-11-06 09:08:32.711925] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.605 [2024-11-06 09:08:32.711935] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.605 [2024-11-06 09:08:32.711942] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.867 [2024-11-06 09:08:32.715468] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.867 [2024-11-06 09:08:32.724599] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.867 [2024-11-06 09:08:32.725266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.867 [2024-11-06 09:08:32.725304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.867 [2024-11-06 09:08:32.725315] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.867 [2024-11-06 09:08:32.725552] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.867 [2024-11-06 09:08:32.725784] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.867 [2024-11-06 09:08:32.725794] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.867 [2024-11-06 09:08:32.725801] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.867 [2024-11-06 09:08:32.729326] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.867 [2024-11-06 09:08:32.738443] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.867 [2024-11-06 09:08:32.739094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.867 [2024-11-06 09:08:32.739132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.867 [2024-11-06 09:08:32.739143] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.867 [2024-11-06 09:08:32.739380] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.867 [2024-11-06 09:08:32.739601] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.867 [2024-11-06 09:08:32.739610] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.867 [2024-11-06 09:08:32.739617] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.867 [2024-11-06 09:08:32.743148] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.867 [2024-11-06 09:08:32.752269] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.867 [2024-11-06 09:08:32.752844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.867 [2024-11-06 09:08:32.752863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.867 [2024-11-06 09:08:32.752871] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.867 [2024-11-06 09:08:32.753090] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.867 [2024-11-06 09:08:32.753312] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.867 [2024-11-06 09:08:32.753320] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.867 [2024-11-06 09:08:32.753327] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.867 [2024-11-06 09:08:32.756846] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.867 [2024-11-06 09:08:32.766170] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.867 [2024-11-06 09:08:32.766827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.867 [2024-11-06 09:08:32.766864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.867 [2024-11-06 09:08:32.766878] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.867 [2024-11-06 09:08:32.767116] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.867 [2024-11-06 09:08:32.767338] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.867 [2024-11-06 09:08:32.767347] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.867 [2024-11-06 09:08:32.767355] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.867 [2024-11-06 09:08:32.770883] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.867 [2024-11-06 09:08:32.780016] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.867 [2024-11-06 09:08:32.780654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.867 [2024-11-06 09:08:32.780693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.867 [2024-11-06 09:08:32.780705] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.867 [2024-11-06 09:08:32.780955] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.867 [2024-11-06 09:08:32.781178] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.867 [2024-11-06 09:08:32.781187] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.867 [2024-11-06 09:08:32.781194] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.867 [2024-11-06 09:08:32.784716] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.867 [2024-11-06 09:08:32.793868] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.867 [2024-11-06 09:08:32.794530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.867 [2024-11-06 09:08:32.794568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.867 [2024-11-06 09:08:32.794579] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.867 [2024-11-06 09:08:32.794827] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.867 [2024-11-06 09:08:32.795049] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.867 [2024-11-06 09:08:32.795059] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.867 [2024-11-06 09:08:32.795067] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.867 [2024-11-06 09:08:32.798591] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.867 [2024-11-06 09:08:32.807723] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.867 [2024-11-06 09:08:32.808295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.867 [2024-11-06 09:08:32.808332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.867 [2024-11-06 09:08:32.808344] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.867 [2024-11-06 09:08:32.808581] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.867 [2024-11-06 09:08:32.808811] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.867 [2024-11-06 09:08:32.808821] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.867 [2024-11-06 09:08:32.808828] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.867 [2024-11-06 09:08:32.812349] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.867 [2024-11-06 09:08:32.821763] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.867 [2024-11-06 09:08:32.822266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.867 [2024-11-06 09:08:32.822303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.867 [2024-11-06 09:08:32.822314] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.867 [2024-11-06 09:08:32.822551] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.867 [2024-11-06 09:08:32.822783] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.867 [2024-11-06 09:08:32.822793] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.867 [2024-11-06 09:08:32.822801] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.867 [2024-11-06 09:08:32.826326] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.867 [2024-11-06 09:08:32.835685] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.867 [2024-11-06 09:08:32.836335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.867 [2024-11-06 09:08:32.836373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.867 [2024-11-06 09:08:32.836385] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.867 [2024-11-06 09:08:32.836621] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.867 [2024-11-06 09:08:32.836852] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.867 [2024-11-06 09:08:32.836862] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.867 [2024-11-06 09:08:32.836870] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.867 [2024-11-06 09:08:32.840396] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.867 [2024-11-06 09:08:32.849531] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.868 [2024-11-06 09:08:32.850081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.868 [2024-11-06 09:08:32.850100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.868 [2024-11-06 09:08:32.850108] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.868 [2024-11-06 09:08:32.850326] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.868 [2024-11-06 09:08:32.850544] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.868 [2024-11-06 09:08:32.850553] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.868 [2024-11-06 09:08:32.850560] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.868 [2024-11-06 09:08:32.854086] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.868 [2024-11-06 09:08:32.863423] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.868 [2024-11-06 09:08:32.864080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.868 [2024-11-06 09:08:32.864118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.868 [2024-11-06 09:08:32.864129] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.868 [2024-11-06 09:08:32.864366] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.868 [2024-11-06 09:08:32.864588] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.868 [2024-11-06 09:08:32.864597] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.868 [2024-11-06 09:08:32.864604] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.868 [2024-11-06 09:08:32.868135] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.868 [2024-11-06 09:08:32.877260] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.868 [2024-11-06 09:08:32.877854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.868 [2024-11-06 09:08:32.877892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.868 [2024-11-06 09:08:32.877904] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.868 [2024-11-06 09:08:32.878144] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.868 [2024-11-06 09:08:32.878365] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.868 [2024-11-06 09:08:32.878376] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.868 [2024-11-06 09:08:32.878383] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.868 [2024-11-06 09:08:32.881913] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.868 [2024-11-06 09:08:32.891055] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.868 [2024-11-06 09:08:32.891673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.868 [2024-11-06 09:08:32.891691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.868 [2024-11-06 09:08:32.891704] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.868 [2024-11-06 09:08:32.891931] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.868 [2024-11-06 09:08:32.892150] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.868 [2024-11-06 09:08:32.892158] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.868 [2024-11-06 09:08:32.892165] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.868 [2024-11-06 09:08:32.895684] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.868 [2024-11-06 09:08:32.904824] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.868 [2024-11-06 09:08:32.905444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.868 [2024-11-06 09:08:32.905481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.868 [2024-11-06 09:08:32.905492] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.868 [2024-11-06 09:08:32.905729] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.868 [2024-11-06 09:08:32.905960] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.868 [2024-11-06 09:08:32.905970] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.868 [2024-11-06 09:08:32.905978] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.868 [2024-11-06 09:08:32.909500] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.868 [2024-11-06 09:08:32.918645] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.868 [2024-11-06 09:08:32.919192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.868 [2024-11-06 09:08:32.919229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.868 [2024-11-06 09:08:32.919241] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.868 [2024-11-06 09:08:32.919482] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.868 [2024-11-06 09:08:32.919703] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.868 [2024-11-06 09:08:32.919711] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.868 [2024-11-06 09:08:32.919719] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.868 [2024-11-06 09:08:32.923250] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.868 [2024-11-06 09:08:32.932589] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.868 [2024-11-06 09:08:32.933129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.868 [2024-11-06 09:08:32.933148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.868 [2024-11-06 09:08:32.933156] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.868 [2024-11-06 09:08:32.933375] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.868 [2024-11-06 09:08:32.933593] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.868 [2024-11-06 09:08:32.933605] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.868 [2024-11-06 09:08:32.933612] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.868 [2024-11-06 09:08:32.937136] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.868 [2024-11-06 09:08:32.946471] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.868 [2024-11-06 09:08:32.947127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.868 [2024-11-06 09:08:32.947165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.868 [2024-11-06 09:08:32.947176] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.868 [2024-11-06 09:08:32.947413] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.868 [2024-11-06 09:08:32.947634] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.868 [2024-11-06 09:08:32.947643] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.868 [2024-11-06 09:08:32.947651] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.868 [2024-11-06 09:08:32.951192] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.868 [2024-11-06 09:08:32.960314] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.868 [2024-11-06 09:08:32.960946] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.868 [2024-11-06 09:08:32.960983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.868 [2024-11-06 09:08:32.960994] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.868 [2024-11-06 09:08:32.961231] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.868 [2024-11-06 09:08:32.961452] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.868 [2024-11-06 09:08:32.961461] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.868 [2024-11-06 09:08:32.961469] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:42.868 [2024-11-06 09:08:32.964996] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:42.868 [2024-11-06 09:08:32.974121] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:42.868 [2024-11-06 09:08:32.974794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:42.868 [2024-11-06 09:08:32.974833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:42.868 [2024-11-06 09:08:32.974845] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:42.868 [2024-11-06 09:08:32.975083] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:42.869 [2024-11-06 09:08:32.975304] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:42.869 [2024-11-06 09:08:32.975313] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:42.869 [2024-11-06 09:08:32.975321] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.131 [2024-11-06 09:08:32.978861] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.131 [2024-11-06 09:08:32.988002] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.131 [2024-11-06 09:08:32.988587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.131 [2024-11-06 09:08:32.988605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.131 [2024-11-06 09:08:32.988613] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.131 [2024-11-06 09:08:32.988838] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.131 [2024-11-06 09:08:32.989057] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.131 [2024-11-06 09:08:32.989065] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.131 [2024-11-06 09:08:32.989073] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.131 [2024-11-06 09:08:32.992584] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.131 [2024-11-06 09:08:33.001913] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.131 [2024-11-06 09:08:33.002484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.131 [2024-11-06 09:08:33.002501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.131 [2024-11-06 09:08:33.002508] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.131 [2024-11-06 09:08:33.002726] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.131 [2024-11-06 09:08:33.002950] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.131 [2024-11-06 09:08:33.002959] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.131 [2024-11-06 09:08:33.002966] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.131 [2024-11-06 09:08:33.006480] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.131 [2024-11-06 09:08:33.015830] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.131 [2024-11-06 09:08:33.016492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.131 [2024-11-06 09:08:33.016529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.131 [2024-11-06 09:08:33.016540] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.131 [2024-11-06 09:08:33.016787] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.131 [2024-11-06 09:08:33.017009] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.131 [2024-11-06 09:08:33.017018] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.131 [2024-11-06 09:08:33.017026] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.131 [2024-11-06 09:08:33.020547] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.131 [2024-11-06 09:08:33.029668] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.131 [2024-11-06 09:08:33.030297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.131 [2024-11-06 09:08:33.030335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.131 [2024-11-06 09:08:33.030346] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.131 [2024-11-06 09:08:33.030583] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.131 [2024-11-06 09:08:33.030813] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.131 [2024-11-06 09:08:33.030823] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.131 [2024-11-06 09:08:33.030830] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.131 [2024-11-06 09:08:33.034353] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.131 [2024-11-06 09:08:33.043487] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.131 [2024-11-06 09:08:33.043938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.131 [2024-11-06 09:08:33.043958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.131 [2024-11-06 09:08:33.043966] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.131 [2024-11-06 09:08:33.044184] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.131 [2024-11-06 09:08:33.044401] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.131 [2024-11-06 09:08:33.044409] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.131 [2024-11-06 09:08:33.044417] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.131 [2024-11-06 09:08:33.047938] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.131 [2024-11-06 09:08:33.057259] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.131 [2024-11-06 09:08:33.057801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.131 [2024-11-06 09:08:33.057825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.131 [2024-11-06 09:08:33.057833] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.131 [2024-11-06 09:08:33.058055] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.131 [2024-11-06 09:08:33.058273] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.131 [2024-11-06 09:08:33.058282] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.131 [2024-11-06 09:08:33.058290] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.131 [2024-11-06 09:08:33.061811] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.131 [2024-11-06 09:08:33.071130] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.131 [2024-11-06 09:08:33.071614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.131 [2024-11-06 09:08:33.071630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.131 [2024-11-06 09:08:33.071638] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.131 [2024-11-06 09:08:33.071865] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.131 [2024-11-06 09:08:33.072084] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.131 [2024-11-06 09:08:33.072092] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.131 [2024-11-06 09:08:33.072099] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.131 [2024-11-06 09:08:33.075642] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.131 [2024-11-06 09:08:33.084975] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.131 [2024-11-06 09:08:33.085594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.131 [2024-11-06 09:08:33.085631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.131 [2024-11-06 09:08:33.085642] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.131 [2024-11-06 09:08:33.085889] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.131 [2024-11-06 09:08:33.086122] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.131 [2024-11-06 09:08:33.086132] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.131 [2024-11-06 09:08:33.086141] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.131 [2024-11-06 09:08:33.089659] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.131 [2024-11-06 09:08:33.098786] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.131 [2024-11-06 09:08:33.099439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.131 [2024-11-06 09:08:33.099477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.131 [2024-11-06 09:08:33.099488] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.131 [2024-11-06 09:08:33.099725] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.131 [2024-11-06 09:08:33.099957] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.131 [2024-11-06 09:08:33.099967] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.131 [2024-11-06 09:08:33.099975] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.131 [2024-11-06 09:08:33.103506] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.131 [2024-11-06 09:08:33.112629] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.131 [2024-11-06 09:08:33.113279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.131 [2024-11-06 09:08:33.113317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.131 [2024-11-06 09:08:33.113328] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.131 [2024-11-06 09:08:33.113565] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.132 [2024-11-06 09:08:33.113797] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.132 [2024-11-06 09:08:33.113816] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.132 [2024-11-06 09:08:33.113824] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.132 [2024-11-06 09:08:33.117344] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.132 [2024-11-06 09:08:33.126476] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.132 [2024-11-06 09:08:33.127118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.132 [2024-11-06 09:08:33.127157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.132 [2024-11-06 09:08:33.127167] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.132 [2024-11-06 09:08:33.127404] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.132 [2024-11-06 09:08:33.127626] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.132 [2024-11-06 09:08:33.127635] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.132 [2024-11-06 09:08:33.127643] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.132 [2024-11-06 09:08:33.131173] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.132 [2024-11-06 09:08:33.140295] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.132 [2024-11-06 09:08:33.140975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.132 [2024-11-06 09:08:33.141013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.132 [2024-11-06 09:08:33.141023] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.132 [2024-11-06 09:08:33.141260] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.132 [2024-11-06 09:08:33.141482] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.132 [2024-11-06 09:08:33.141491] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.132 [2024-11-06 09:08:33.141499] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.132 [2024-11-06 09:08:33.145031] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.132 [2024-11-06 09:08:33.154153] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.132 [2024-11-06 09:08:33.154744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.132 [2024-11-06 09:08:33.154797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.132 [2024-11-06 09:08:33.154809] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.132 [2024-11-06 09:08:33.155046] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.132 [2024-11-06 09:08:33.155267] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.132 [2024-11-06 09:08:33.155276] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.132 [2024-11-06 09:08:33.155283] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.132 [2024-11-06 09:08:33.158813] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.132 [2024-11-06 09:08:33.167930] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.132 [2024-11-06 09:08:33.168507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.132 [2024-11-06 09:08:33.168525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.132 [2024-11-06 09:08:33.168534] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.132 [2024-11-06 09:08:33.168758] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.132 [2024-11-06 09:08:33.168976] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.132 [2024-11-06 09:08:33.168984] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.132 [2024-11-06 09:08:33.168991] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.132 [2024-11-06 09:08:33.172505] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.132 [2024-11-06 09:08:33.181835] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.132 [2024-11-06 09:08:33.182398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.132 [2024-11-06 09:08:33.182414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.132 [2024-11-06 09:08:33.182421] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.132 [2024-11-06 09:08:33.182639] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.132 [2024-11-06 09:08:33.182863] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.132 [2024-11-06 09:08:33.182871] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.132 [2024-11-06 09:08:33.182878] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.132 5698.20 IOPS, 22.26 MiB/s [2024-11-06T08:08:33.245Z] [2024-11-06 09:08:33.188051] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.132 [2024-11-06 09:08:33.195717] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.132 [2024-11-06 09:08:33.196333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.132 [2024-11-06 09:08:33.196371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.132 [2024-11-06 09:08:33.196382] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.132 [2024-11-06 09:08:33.196618] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.132 [2024-11-06 09:08:33.196850] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.132 [2024-11-06 09:08:33.196860] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.132 [2024-11-06 09:08:33.196868] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.132 [2024-11-06 09:08:33.200390] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.132 [2024-11-06 09:08:33.209543] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.132 [2024-11-06 09:08:33.210206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.132 [2024-11-06 09:08:33.210244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.132 [2024-11-06 09:08:33.210255] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.132 [2024-11-06 09:08:33.210492] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.132 [2024-11-06 09:08:33.210713] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.132 [2024-11-06 09:08:33.210722] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.132 [2024-11-06 09:08:33.210730] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.132 [2024-11-06 09:08:33.214266] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.132 [2024-11-06 09:08:33.223408] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.132 [2024-11-06 09:08:33.223947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.132 [2024-11-06 09:08:33.223967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.132 [2024-11-06 09:08:33.223975] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.132 [2024-11-06 09:08:33.224193] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.132 [2024-11-06 09:08:33.224411] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.132 [2024-11-06 09:08:33.224419] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.132 [2024-11-06 09:08:33.224426] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.132 [2024-11-06 09:08:33.227950] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.132 [2024-11-06 09:08:33.237286] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.132 [2024-11-06 09:08:33.237850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.132 [2024-11-06 09:08:33.237866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.132 [2024-11-06 09:08:33.237873] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.132 [2024-11-06 09:08:33.238091] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.132 [2024-11-06 09:08:33.238308] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.132 [2024-11-06 09:08:33.238317] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.132 [2024-11-06 09:08:33.238324] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.394 [2024-11-06 09:08:33.241848] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.394 [2024-11-06 09:08:33.251207] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.394 [2024-11-06 09:08:33.251768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.394 [2024-11-06 09:08:33.251784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.394 [2024-11-06 09:08:33.251792] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.394 [2024-11-06 09:08:33.252014] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.394 [2024-11-06 09:08:33.252231] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.394 [2024-11-06 09:08:33.252239] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.394 [2024-11-06 09:08:33.252246] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.394 [2024-11-06 09:08:33.255774] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.394 [2024-11-06 09:08:33.265125] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.394 [2024-11-06 09:08:33.265768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.394 [2024-11-06 09:08:33.265807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.394 [2024-11-06 09:08:33.265819] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.394 [2024-11-06 09:08:33.266059] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.394 [2024-11-06 09:08:33.266281] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.394 [2024-11-06 09:08:33.266290] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.394 [2024-11-06 09:08:33.266298] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.394 [2024-11-06 09:08:33.269827] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.394 [2024-11-06 09:08:33.278950] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.394 [2024-11-06 09:08:33.279532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.394 [2024-11-06 09:08:33.279551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.394 [2024-11-06 09:08:33.279559] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.394 [2024-11-06 09:08:33.279784] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.394 [2024-11-06 09:08:33.280002] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.394 [2024-11-06 09:08:33.280010] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.394 [2024-11-06 09:08:33.280017] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.394 [2024-11-06 09:08:33.283529] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.394 [2024-11-06 09:08:33.292868] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.394 [2024-11-06 09:08:33.293472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.394 [2024-11-06 09:08:33.293510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.394 [2024-11-06 09:08:33.293521] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.394 [2024-11-06 09:08:33.293766] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.395 [2024-11-06 09:08:33.293989] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.395 [2024-11-06 09:08:33.294003] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.395 [2024-11-06 09:08:33.294011] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.395 [2024-11-06 09:08:33.297534] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.395 [2024-11-06 09:08:33.306674] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.395 [2024-11-06 09:08:33.307354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.395 [2024-11-06 09:08:33.307392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.395 [2024-11-06 09:08:33.307403] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.395 [2024-11-06 09:08:33.307640] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.395 [2024-11-06 09:08:33.307871] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.395 [2024-11-06 09:08:33.307881] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.395 [2024-11-06 09:08:33.307888] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.395 [2024-11-06 09:08:33.311409] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.395 [2024-11-06 09:08:33.320534] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.395 [2024-11-06 09:08:33.321160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.395 [2024-11-06 09:08:33.321198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.395 [2024-11-06 09:08:33.321208] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.395 [2024-11-06 09:08:33.321445] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.395 [2024-11-06 09:08:33.321667] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.395 [2024-11-06 09:08:33.321676] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.395 [2024-11-06 09:08:33.321683] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.395 [2024-11-06 09:08:33.325212] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.395 [2024-11-06 09:08:33.334347] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.395 [2024-11-06 09:08:33.335040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.395 [2024-11-06 09:08:33.335078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.395 [2024-11-06 09:08:33.335089] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.395 [2024-11-06 09:08:33.335325] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.395 [2024-11-06 09:08:33.335547] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.395 [2024-11-06 09:08:33.335556] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.395 [2024-11-06 09:08:33.335564] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.395 [2024-11-06 09:08:33.339100] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.395 [2024-11-06 09:08:33.348224] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.395 [2024-11-06 09:08:33.348948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.395 [2024-11-06 09:08:33.348986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.395 [2024-11-06 09:08:33.348997] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.395 [2024-11-06 09:08:33.349233] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.395 [2024-11-06 09:08:33.349455] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.395 [2024-11-06 09:08:33.349464] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.395 [2024-11-06 09:08:33.349472] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.395 [2024-11-06 09:08:33.353001] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.395 [2024-11-06 09:08:33.362131] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.395 [2024-11-06 09:08:33.362724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.395 [2024-11-06 09:08:33.362743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.395 [2024-11-06 09:08:33.362762] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.395 [2024-11-06 09:08:33.362989] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.395 [2024-11-06 09:08:33.363209] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.395 [2024-11-06 09:08:33.363217] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.395 [2024-11-06 09:08:33.363224] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.395 [2024-11-06 09:08:33.366748] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.395 [2024-11-06 09:08:33.376084] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.395 [2024-11-06 09:08:33.376646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.395 [2024-11-06 09:08:33.376662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.395 [2024-11-06 09:08:33.376670] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.395 [2024-11-06 09:08:33.376894] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.395 [2024-11-06 09:08:33.377112] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.395 [2024-11-06 09:08:33.377120] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.395 [2024-11-06 09:08:33.377127] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.395 [2024-11-06 09:08:33.380644] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.395 [2024-11-06 09:08:33.389997] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.395 [2024-11-06 09:08:33.390655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.395 [2024-11-06 09:08:33.390693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.395 [2024-11-06 09:08:33.390705] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.395 [2024-11-06 09:08:33.390953] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.395 [2024-11-06 09:08:33.391175] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.395 [2024-11-06 09:08:33.391184] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.395 [2024-11-06 09:08:33.391192] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.395 [2024-11-06 09:08:33.394713] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.395 [2024-11-06 09:08:33.403891] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.395 [2024-11-06 09:08:33.404548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.395 [2024-11-06 09:08:33.404585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.395 [2024-11-06 09:08:33.404596] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.395 [2024-11-06 09:08:33.404843] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.395 [2024-11-06 09:08:33.405065] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.395 [2024-11-06 09:08:33.405075] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.395 [2024-11-06 09:08:33.405082] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.395 [2024-11-06 09:08:33.408603] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.395 [2024-11-06 09:08:33.417725] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.395 [2024-11-06 09:08:33.418398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.395 [2024-11-06 09:08:33.418436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.395 [2024-11-06 09:08:33.418446] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.395 [2024-11-06 09:08:33.418683] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.395 [2024-11-06 09:08:33.418914] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.395 [2024-11-06 09:08:33.418924] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.395 [2024-11-06 09:08:33.418932] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.395 [2024-11-06 09:08:33.422454] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.395 [2024-11-06 09:08:33.431575] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.395 [2024-11-06 09:08:33.432142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.395 [2024-11-06 09:08:33.432161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.395 [2024-11-06 09:08:33.432169] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.395 [2024-11-06 09:08:33.432391] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.396 [2024-11-06 09:08:33.432609] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.396 [2024-11-06 09:08:33.432617] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.396 [2024-11-06 09:08:33.432624] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.396 [2024-11-06 09:08:33.436147] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.396 [2024-11-06 09:08:33.445470] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.396 [2024-11-06 09:08:33.446022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.396 [2024-11-06 09:08:33.446060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.396 [2024-11-06 09:08:33.446071] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.396 [2024-11-06 09:08:33.446308] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.396 [2024-11-06 09:08:33.446530] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.396 [2024-11-06 09:08:33.446538] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.396 [2024-11-06 09:08:33.446546] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.396 [2024-11-06 09:08:33.450075] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.396 [2024-11-06 09:08:33.459407] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.396 [2024-11-06 09:08:33.460038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.396 [2024-11-06 09:08:33.460076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.396 [2024-11-06 09:08:33.460087] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.396 [2024-11-06 09:08:33.460323] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.396 [2024-11-06 09:08:33.460545] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.396 [2024-11-06 09:08:33.460554] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.396 [2024-11-06 09:08:33.460561] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.396 [2024-11-06 09:08:33.464091] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.396 [2024-11-06 09:08:33.473214] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.396 [2024-11-06 09:08:33.473877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.396 [2024-11-06 09:08:33.473915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.396 [2024-11-06 09:08:33.473925] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.396 [2024-11-06 09:08:33.474162] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.396 [2024-11-06 09:08:33.474384] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.396 [2024-11-06 09:08:33.474397] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.396 [2024-11-06 09:08:33.474405] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.396 [2024-11-06 09:08:33.477936] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.396 [2024-11-06 09:08:33.487063] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.396 [2024-11-06 09:08:33.487744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.396 [2024-11-06 09:08:33.487789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.396 [2024-11-06 09:08:33.487800] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.396 [2024-11-06 09:08:33.488037] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.396 [2024-11-06 09:08:33.488258] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.396 [2024-11-06 09:08:33.488267] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.396 [2024-11-06 09:08:33.488275] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.396 [2024-11-06 09:08:33.491801] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.396 [2024-11-06 09:08:33.500936] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.396 [2024-11-06 09:08:33.501608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.396 [2024-11-06 09:08:33.501646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.396 [2024-11-06 09:08:33.501657] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.396 [2024-11-06 09:08:33.501904] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.396 [2024-11-06 09:08:33.502126] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.396 [2024-11-06 09:08:33.502134] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.396 [2024-11-06 09:08:33.502142] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.663 [2024-11-06 09:08:33.505663] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.663 [2024-11-06 09:08:33.514792] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.663 [2024-11-06 09:08:33.515351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.663 [2024-11-06 09:08:33.515389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.663 [2024-11-06 09:08:33.515402] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.663 [2024-11-06 09:08:33.515640] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.663 [2024-11-06 09:08:33.515871] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.663 [2024-11-06 09:08:33.515881] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.663 [2024-11-06 09:08:33.515889] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.663 [2024-11-06 09:08:33.519416] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.663 [2024-11-06 09:08:33.528756] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.663 [2024-11-06 09:08:33.529386] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.663 [2024-11-06 09:08:33.529424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.663 [2024-11-06 09:08:33.529435] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.663 [2024-11-06 09:08:33.529672] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.663 [2024-11-06 09:08:33.529903] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.663 [2024-11-06 09:08:33.529913] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.663 [2024-11-06 09:08:33.529920] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.663 [2024-11-06 09:08:33.533443] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.663 [2024-11-06 09:08:33.542574] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.663 [2024-11-06 09:08:33.543221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.663 [2024-11-06 09:08:33.543259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.663 [2024-11-06 09:08:33.543270] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.664 [2024-11-06 09:08:33.543507] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.664 [2024-11-06 09:08:33.543729] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.664 [2024-11-06 09:08:33.543738] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.664 [2024-11-06 09:08:33.543756] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.664 [2024-11-06 09:08:33.547278] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.664 [2024-11-06 09:08:33.556409] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.664 [2024-11-06 09:08:33.557104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.664 [2024-11-06 09:08:33.557143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.664 [2024-11-06 09:08:33.557154] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.664 [2024-11-06 09:08:33.557391] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.664 [2024-11-06 09:08:33.557613] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.664 [2024-11-06 09:08:33.557622] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.664 [2024-11-06 09:08:33.557629] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.664 [2024-11-06 09:08:33.561158] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.664 [2024-11-06 09:08:33.570282] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.664 [2024-11-06 09:08:33.570973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.664 [2024-11-06 09:08:33.571015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.664 [2024-11-06 09:08:33.571026] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.664 [2024-11-06 09:08:33.571263] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.664 [2024-11-06 09:08:33.571485] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.664 [2024-11-06 09:08:33.571493] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.664 [2024-11-06 09:08:33.571501] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.664 [2024-11-06 09:08:33.575032] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.664 [2024-11-06 09:08:33.584157] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.664 [2024-11-06 09:08:33.584833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.664 [2024-11-06 09:08:33.584871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.664 [2024-11-06 09:08:33.584882] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.664 [2024-11-06 09:08:33.585119] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.664 [2024-11-06 09:08:33.585340] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.664 [2024-11-06 09:08:33.585349] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.664 [2024-11-06 09:08:33.585357] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.664 [2024-11-06 09:08:33.588899] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.664 [2024-11-06 09:08:33.598028] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.664 [2024-11-06 09:08:33.598701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.664 [2024-11-06 09:08:33.598738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.664 [2024-11-06 09:08:33.598759] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.664 [2024-11-06 09:08:33.599000] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.664 [2024-11-06 09:08:33.599221] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.664 [2024-11-06 09:08:33.599230] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.664 [2024-11-06 09:08:33.599238] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.664 [2024-11-06 09:08:33.602771] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.664 [2024-11-06 09:08:33.611895] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.664 [2024-11-06 09:08:33.612550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.664 [2024-11-06 09:08:33.612588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.664 [2024-11-06 09:08:33.612599] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.664 [2024-11-06 09:08:33.612849] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.664 [2024-11-06 09:08:33.613071] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.664 [2024-11-06 09:08:33.613080] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.664 [2024-11-06 09:08:33.613088] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.664 [2024-11-06 09:08:33.616609] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.664 [2024-11-06 09:08:33.625731] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.664 [2024-11-06 09:08:33.626359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.664 [2024-11-06 09:08:33.626397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.664 [2024-11-06 09:08:33.626408] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.664 [2024-11-06 09:08:33.626644] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.664 [2024-11-06 09:08:33.626875] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.664 [2024-11-06 09:08:33.626885] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.664 [2024-11-06 09:08:33.626893] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.664 [2024-11-06 09:08:33.630413] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.664 [2024-11-06 09:08:33.639535] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.664 [2024-11-06 09:08:33.640191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.664 [2024-11-06 09:08:33.640229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.664 [2024-11-06 09:08:33.640240] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.664 [2024-11-06 09:08:33.640477] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.664 [2024-11-06 09:08:33.640699] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.664 [2024-11-06 09:08:33.640708] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.664 [2024-11-06 09:08:33.640715] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.664 [2024-11-06 09:08:33.644245] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.664 [2024-11-06 09:08:33.653371] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.664 [2024-11-06 09:08:33.654065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.664 [2024-11-06 09:08:33.654104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.664 [2024-11-06 09:08:33.654114] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.664 [2024-11-06 09:08:33.654351] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.664 [2024-11-06 09:08:33.654572] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.664 [2024-11-06 09:08:33.654585] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.664 [2024-11-06 09:08:33.654593] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.664 [2024-11-06 09:08:33.658124] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.664 [2024-11-06 09:08:33.667246] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.664 [2024-11-06 09:08:33.667948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.664 [2024-11-06 09:08:33.667986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.664 [2024-11-06 09:08:33.667997] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.664 [2024-11-06 09:08:33.668234] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.664 [2024-11-06 09:08:33.668455] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.664 [2024-11-06 09:08:33.668463] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.664 [2024-11-06 09:08:33.668472] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.664 [2024-11-06 09:08:33.672001] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.664 [2024-11-06 09:08:33.681126] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.664 [2024-11-06 09:08:33.681826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.664 [2024-11-06 09:08:33.681864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.664 [2024-11-06 09:08:33.681875] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.665 [2024-11-06 09:08:33.682112] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.665 [2024-11-06 09:08:33.682332] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.665 [2024-11-06 09:08:33.682341] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.665 [2024-11-06 09:08:33.682349] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.665 [2024-11-06 09:08:33.685879] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.665 [2024-11-06 09:08:33.695012] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.665 [2024-11-06 09:08:33.695589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.665 [2024-11-06 09:08:33.695608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.665 [2024-11-06 09:08:33.695616] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.665 [2024-11-06 09:08:33.695840] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.665 [2024-11-06 09:08:33.696058] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.665 [2024-11-06 09:08:33.696066] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.665 [2024-11-06 09:08:33.696073] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.665 [2024-11-06 09:08:33.699595] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.665 [2024-11-06 09:08:33.708938] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.665 [2024-11-06 09:08:33.709581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.665 [2024-11-06 09:08:33.709619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.665 [2024-11-06 09:08:33.709630] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.665 [2024-11-06 09:08:33.709875] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.665 [2024-11-06 09:08:33.710098] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.665 [2024-11-06 09:08:33.710107] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.665 [2024-11-06 09:08:33.710114] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.665 [2024-11-06 09:08:33.713635] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.665 [2024-11-06 09:08:33.722771] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.665 [2024-11-06 09:08:33.723344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.665 [2024-11-06 09:08:33.723382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.665 [2024-11-06 09:08:33.723393] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.665 [2024-11-06 09:08:33.723630] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.665 [2024-11-06 09:08:33.723861] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.665 [2024-11-06 09:08:33.723871] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.665 [2024-11-06 09:08:33.723879] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.665 [2024-11-06 09:08:33.727401] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.665 [2024-11-06 09:08:33.736726] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.665 [2024-11-06 09:08:33.737376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.665 [2024-11-06 09:08:33.737414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.665 [2024-11-06 09:08:33.737424] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.665 [2024-11-06 09:08:33.737661] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.665 [2024-11-06 09:08:33.737892] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.665 [2024-11-06 09:08:33.737902] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.665 [2024-11-06 09:08:33.737910] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.665 [2024-11-06 09:08:33.741432] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.665 [2024-11-06 09:08:33.750559] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.665 [2024-11-06 09:08:33.751200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.665 [2024-11-06 09:08:33.751242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.665 [2024-11-06 09:08:33.751253] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.665 [2024-11-06 09:08:33.751490] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.665 [2024-11-06 09:08:33.751712] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.665 [2024-11-06 09:08:33.751722] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.665 [2024-11-06 09:08:33.751731] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.665 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh: line 35: 2072404 Killed "${NVMF_APP[@]}" "$@" 00:32:43.665 [2024-11-06 09:08:33.755261] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@36 -- # tgt_init 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@724 -- # xtrace_disable 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:43.665 [2024-11-06 09:08:33.764393] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@507 -- # nvmfpid=2073990 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@508 -- # waitforlisten 2073990 00:32:43.665 [2024-11-06 09:08:33.765047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.665 [2024-11-06 09:08:33.765085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.665 [2024-11-06 09:08:33.765096] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:32:43.665 [2024-11-06 09:08:33.765333] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@831 -- # '[' -z 2073990 ']' 00:32:43.665 [2024-11-06 09:08:33.765555] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.665 [2024-11-06 09:08:33.765565] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.665 [2024-11-06 09:08:33.765573] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@836 -- # local max_retries=100 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:32:43.665 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@840 -- # xtrace_disable 00:32:43.665 09:08:33 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:43.952 [2024-11-06 09:08:33.769105] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.952 [2024-11-06 09:08:33.778274] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.952 [2024-11-06 09:08:33.778755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.952 [2024-11-06 09:08:33.778781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.952 [2024-11-06 09:08:33.778795] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.952 [2024-11-06 09:08:33.779014] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.952 [2024-11-06 09:08:33.779234] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.952 [2024-11-06 09:08:33.779243] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.952 [2024-11-06 09:08:33.779250] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.952 [2024-11-06 09:08:33.782773] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.952 [2024-11-06 09:08:33.792122] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.952 [2024-11-06 09:08:33.792774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.952 [2024-11-06 09:08:33.792813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.952 [2024-11-06 09:08:33.792826] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.952 [2024-11-06 09:08:33.793067] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.952 [2024-11-06 09:08:33.793288] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.952 [2024-11-06 09:08:33.793298] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.952 [2024-11-06 09:08:33.793306] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.952 [2024-11-06 09:08:33.796840] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.952 [2024-11-06 09:08:33.805986] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.952 [2024-11-06 09:08:33.806631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.952 [2024-11-06 09:08:33.806669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.952 [2024-11-06 09:08:33.806680] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.952 [2024-11-06 09:08:33.806926] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.952 [2024-11-06 09:08:33.807148] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.952 [2024-11-06 09:08:33.807157] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.952 [2024-11-06 09:08:33.807166] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.952 [2024-11-06 09:08:33.810687] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.952 [2024-11-06 09:08:33.819825] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.952 [2024-11-06 09:08:33.820474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.952 [2024-11-06 09:08:33.820511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.952 [2024-11-06 09:08:33.820529] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.952 [2024-11-06 09:08:33.820773] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.952 [2024-11-06 09:08:33.820995] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.952 [2024-11-06 09:08:33.821004] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.952 [2024-11-06 09:08:33.821012] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.952 [2024-11-06 09:08:33.822944] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:32:43.953 [2024-11-06 09:08:33.823014] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:32:43.953 [2024-11-06 09:08:33.824533] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.953 [2024-11-06 09:08:33.833664] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.953 [2024-11-06 09:08:33.834348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.953 [2024-11-06 09:08:33.834386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.953 [2024-11-06 09:08:33.834397] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.953 [2024-11-06 09:08:33.834634] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.953 [2024-11-06 09:08:33.834864] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.953 [2024-11-06 09:08:33.834874] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.953 [2024-11-06 09:08:33.834883] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.953 [2024-11-06 09:08:33.838403] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.953 [2024-11-06 09:08:33.847530] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.953 [2024-11-06 09:08:33.848195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.953 [2024-11-06 09:08:33.848233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.953 [2024-11-06 09:08:33.848244] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.953 [2024-11-06 09:08:33.848481] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.953 [2024-11-06 09:08:33.848703] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.953 [2024-11-06 09:08:33.848713] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.953 [2024-11-06 09:08:33.848721] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.953 [2024-11-06 09:08:33.852327] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.953 [2024-11-06 09:08:33.861468] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.953 [2024-11-06 09:08:33.862142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.953 [2024-11-06 09:08:33.862181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.953 [2024-11-06 09:08:33.862196] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.953 [2024-11-06 09:08:33.862434] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.953 [2024-11-06 09:08:33.862655] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.953 [2024-11-06 09:08:33.862665] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.953 [2024-11-06 09:08:33.862673] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.953 [2024-11-06 09:08:33.866202] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.953 [2024-11-06 09:08:33.875334] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.953 [2024-11-06 09:08:33.875896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.953 [2024-11-06 09:08:33.875916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.953 [2024-11-06 09:08:33.875924] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.953 [2024-11-06 09:08:33.876143] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.953 [2024-11-06 09:08:33.876360] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.953 [2024-11-06 09:08:33.876369] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.953 [2024-11-06 09:08:33.876376] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.953 [2024-11-06 09:08:33.879898] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.953 [2024-11-06 09:08:33.889243] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.953 [2024-11-06 09:08:33.889820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.953 [2024-11-06 09:08:33.889836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.953 [2024-11-06 09:08:33.889844] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.953 [2024-11-06 09:08:33.890062] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.953 [2024-11-06 09:08:33.890279] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.953 [2024-11-06 09:08:33.890288] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.953 [2024-11-06 09:08:33.890295] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.953 [2024-11-06 09:08:33.893816] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.953 [2024-11-06 09:08:33.903154] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.953 [2024-11-06 09:08:33.903702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.953 [2024-11-06 09:08:33.903718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.953 [2024-11-06 09:08:33.903725] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.953 [2024-11-06 09:08:33.903948] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.953 [2024-11-06 09:08:33.904170] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.953 [2024-11-06 09:08:33.904178] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.953 [2024-11-06 09:08:33.904186] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.953 [2024-11-06 09:08:33.907699] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.953 [2024-11-06 09:08:33.912650] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:32:43.953 [2024-11-06 09:08:33.917032] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.953 [2024-11-06 09:08:33.917568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.953 [2024-11-06 09:08:33.917584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.953 [2024-11-06 09:08:33.917592] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.953 [2024-11-06 09:08:33.917814] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.953 [2024-11-06 09:08:33.918032] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.953 [2024-11-06 09:08:33.918040] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.953 [2024-11-06 09:08:33.918048] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.953 [2024-11-06 09:08:33.921566] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.953 [2024-11-06 09:08:33.930906] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.953 [2024-11-06 09:08:33.931446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.953 [2024-11-06 09:08:33.931461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.953 [2024-11-06 09:08:33.931469] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.953 [2024-11-06 09:08:33.931687] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.953 [2024-11-06 09:08:33.931912] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.953 [2024-11-06 09:08:33.931922] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.953 [2024-11-06 09:08:33.931929] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.953 [2024-11-06 09:08:33.935445] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.953 [2024-11-06 09:08:33.941919] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:32:43.954 [2024-11-06 09:08:33.941942] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:32:43.954 [2024-11-06 09:08:33.941950] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:32:43.954 [2024-11-06 09:08:33.941956] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:32:43.954 [2024-11-06 09:08:33.941961] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:32:43.954 [2024-11-06 09:08:33.943054] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:32:43.954 [2024-11-06 09:08:33.943294] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:32:43.954 [2024-11-06 09:08:33.943295] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:32:43.954 [2024-11-06 09:08:33.944782] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.954 [2024-11-06 09:08:33.945506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.954 [2024-11-06 09:08:33.945547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.954 [2024-11-06 09:08:33.945559] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.954 [2024-11-06 09:08:33.945811] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.954 [2024-11-06 09:08:33.946035] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.954 [2024-11-06 09:08:33.946044] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.954 [2024-11-06 09:08:33.946053] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.954 [2024-11-06 09:08:33.949577] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.954 [2024-11-06 09:08:33.958718] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.954 [2024-11-06 09:08:33.959326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.954 [2024-11-06 09:08:33.959366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.954 [2024-11-06 09:08:33.959378] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.954 [2024-11-06 09:08:33.959618] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.954 [2024-11-06 09:08:33.959847] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.954 [2024-11-06 09:08:33.959858] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.954 [2024-11-06 09:08:33.959866] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.954 [2024-11-06 09:08:33.963389] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.954 [2024-11-06 09:08:33.972523] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.954 [2024-11-06 09:08:33.973127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.954 [2024-11-06 09:08:33.973146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.954 [2024-11-06 09:08:33.973155] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.954 [2024-11-06 09:08:33.973373] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.954 [2024-11-06 09:08:33.973590] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.954 [2024-11-06 09:08:33.973599] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.954 [2024-11-06 09:08:33.973607] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.954 [2024-11-06 09:08:33.977130] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.954 [2024-11-06 09:08:33.986475] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.954 [2024-11-06 09:08:33.987154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.954 [2024-11-06 09:08:33.987195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.954 [2024-11-06 09:08:33.987213] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.954 [2024-11-06 09:08:33.987453] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.954 [2024-11-06 09:08:33.987676] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.954 [2024-11-06 09:08:33.987685] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.954 [2024-11-06 09:08:33.987693] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.954 [2024-11-06 09:08:33.991241] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.954 [2024-11-06 09:08:34.000372] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.954 [2024-11-06 09:08:34.001072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.954 [2024-11-06 09:08:34.001111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.954 [2024-11-06 09:08:34.001122] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.954 [2024-11-06 09:08:34.001359] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.954 [2024-11-06 09:08:34.001581] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.954 [2024-11-06 09:08:34.001591] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.954 [2024-11-06 09:08:34.001599] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.954 [2024-11-06 09:08:34.005145] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.954 [2024-11-06 09:08:34.014276] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.954 [2024-11-06 09:08:34.014866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.954 [2024-11-06 09:08:34.014904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.954 [2024-11-06 09:08:34.014917] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.954 [2024-11-06 09:08:34.015158] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.954 [2024-11-06 09:08:34.015379] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.954 [2024-11-06 09:08:34.015388] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.954 [2024-11-06 09:08:34.015396] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.954 [2024-11-06 09:08:34.018927] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.954 [2024-11-06 09:08:34.028055] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.954 [2024-11-06 09:08:34.028624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.954 [2024-11-06 09:08:34.028643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.954 [2024-11-06 09:08:34.028651] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.954 [2024-11-06 09:08:34.028875] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.954 [2024-11-06 09:08:34.029099] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.954 [2024-11-06 09:08:34.029107] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.954 [2024-11-06 09:08:34.029114] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.954 [2024-11-06 09:08:34.032629] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:43.954 [2024-11-06 09:08:34.041962] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:43.954 [2024-11-06 09:08:34.042638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:43.954 [2024-11-06 09:08:34.042677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:43.954 [2024-11-06 09:08:34.042688] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:43.954 [2024-11-06 09:08:34.042933] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:43.954 [2024-11-06 09:08:34.043156] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:43.954 [2024-11-06 09:08:34.043166] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:43.955 [2024-11-06 09:08:34.043174] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:43.955 [2024-11-06 09:08:34.046695] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.241 [2024-11-06 09:08:34.055829] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.241 [2024-11-06 09:08:34.056389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.241 [2024-11-06 09:08:34.056408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.241 [2024-11-06 09:08:34.056416] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.241 [2024-11-06 09:08:34.056633] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.241 [2024-11-06 09:08:34.056857] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.241 [2024-11-06 09:08:34.056875] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.241 [2024-11-06 09:08:34.056883] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.241 [2024-11-06 09:08:34.060402] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.241 [2024-11-06 09:08:34.069738] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.241 [2024-11-06 09:08:34.070313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.241 [2024-11-06 09:08:34.070330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.241 [2024-11-06 09:08:34.070337] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.241 [2024-11-06 09:08:34.070554] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.241 [2024-11-06 09:08:34.070778] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.241 [2024-11-06 09:08:34.070788] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.241 [2024-11-06 09:08:34.070800] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.241 [2024-11-06 09:08:34.074318] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.241 [2024-11-06 09:08:34.083660] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.241 [2024-11-06 09:08:34.084312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.241 [2024-11-06 09:08:34.084350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.241 [2024-11-06 09:08:34.084361] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.241 [2024-11-06 09:08:34.084599] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.241 [2024-11-06 09:08:34.084827] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.241 [2024-11-06 09:08:34.084837] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.241 [2024-11-06 09:08:34.084845] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.241 [2024-11-06 09:08:34.088367] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.241 [2024-11-06 09:08:34.097506] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.241 [2024-11-06 09:08:34.098031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.241 [2024-11-06 09:08:34.098051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.241 [2024-11-06 09:08:34.098059] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.241 [2024-11-06 09:08:34.098277] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.241 [2024-11-06 09:08:34.098494] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.241 [2024-11-06 09:08:34.098502] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.241 [2024-11-06 09:08:34.098510] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.242 [2024-11-06 09:08:34.102132] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.242 [2024-11-06 09:08:34.111467] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.242 [2024-11-06 09:08:34.112000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.242 [2024-11-06 09:08:34.112039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.242 [2024-11-06 09:08:34.112050] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.242 [2024-11-06 09:08:34.112286] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.242 [2024-11-06 09:08:34.112508] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.242 [2024-11-06 09:08:34.112517] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.242 [2024-11-06 09:08:34.112525] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.242 [2024-11-06 09:08:34.116054] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.242 [2024-11-06 09:08:34.125394] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.242 [2024-11-06 09:08:34.125983] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.242 [2024-11-06 09:08:34.126021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.242 [2024-11-06 09:08:34.126033] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.242 [2024-11-06 09:08:34.126272] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.242 [2024-11-06 09:08:34.126493] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.242 [2024-11-06 09:08:34.126502] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.242 [2024-11-06 09:08:34.126510] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.242 [2024-11-06 09:08:34.130041] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.242 [2024-11-06 09:08:34.139169] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.242 [2024-11-06 09:08:34.139848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.242 [2024-11-06 09:08:34.139887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.242 [2024-11-06 09:08:34.139899] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.242 [2024-11-06 09:08:34.140140] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.242 [2024-11-06 09:08:34.140362] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.242 [2024-11-06 09:08:34.140371] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.242 [2024-11-06 09:08:34.140379] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.242 [2024-11-06 09:08:34.143911] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.242 [2024-11-06 09:08:34.153037] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.242 [2024-11-06 09:08:34.153591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.242 [2024-11-06 09:08:34.153609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.242 [2024-11-06 09:08:34.153617] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.242 [2024-11-06 09:08:34.153842] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.242 [2024-11-06 09:08:34.154060] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.242 [2024-11-06 09:08:34.154069] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.242 [2024-11-06 09:08:34.154076] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.242 [2024-11-06 09:08:34.157595] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.242 [2024-11-06 09:08:34.166926] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.242 [2024-11-06 09:08:34.167601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.242 [2024-11-06 09:08:34.167640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.242 [2024-11-06 09:08:34.167655] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.242 [2024-11-06 09:08:34.167900] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.242 [2024-11-06 09:08:34.168122] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.242 [2024-11-06 09:08:34.168131] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.242 [2024-11-06 09:08:34.168138] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.242 [2024-11-06 09:08:34.171658] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.242 [2024-11-06 09:08:34.180805] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.242 [2024-11-06 09:08:34.181353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.242 [2024-11-06 09:08:34.181373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.242 [2024-11-06 09:08:34.181381] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.242 [2024-11-06 09:08:34.181599] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.242 [2024-11-06 09:08:34.181823] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.242 [2024-11-06 09:08:34.181832] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.242 [2024-11-06 09:08:34.181839] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.242 [2024-11-06 09:08:34.185367] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.242 4748.50 IOPS, 18.55 MiB/s [2024-11-06T08:08:34.355Z] [2024-11-06 09:08:34.194714] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.242 [2024-11-06 09:08:34.195394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.242 [2024-11-06 09:08:34.195432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.242 [2024-11-06 09:08:34.195443] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.242 [2024-11-06 09:08:34.195680] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.242 [2024-11-06 09:08:34.195910] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.242 [2024-11-06 09:08:34.195920] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.242 [2024-11-06 09:08:34.195927] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.242 [2024-11-06 09:08:34.199448] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.242 [2024-11-06 09:08:34.208590] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.242 [2024-11-06 09:08:34.209157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.242 [2024-11-06 09:08:34.209177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.242 [2024-11-06 09:08:34.209185] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.242 [2024-11-06 09:08:34.209403] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.242 [2024-11-06 09:08:34.209625] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.242 [2024-11-06 09:08:34.209634] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.242 [2024-11-06 09:08:34.209641] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.242 [2024-11-06 09:08:34.213166] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.242 [2024-11-06 09:08:34.222497] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.242 [2024-11-06 09:08:34.223117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.242 [2024-11-06 09:08:34.223156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.242 [2024-11-06 09:08:34.223167] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.242 [2024-11-06 09:08:34.223404] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.242 [2024-11-06 09:08:34.223625] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.242 [2024-11-06 09:08:34.223634] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.242 [2024-11-06 09:08:34.223642] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.242 [2024-11-06 09:08:34.227170] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.242 [2024-11-06 09:08:34.236298] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.242 [2024-11-06 09:08:34.236847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.242 [2024-11-06 09:08:34.236868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.242 [2024-11-06 09:08:34.236876] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.242 [2024-11-06 09:08:34.237094] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.243 [2024-11-06 09:08:34.237311] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.243 [2024-11-06 09:08:34.237320] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.243 [2024-11-06 09:08:34.237328] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.243 [2024-11-06 09:08:34.240849] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.243 [2024-11-06 09:08:34.250183] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.243 [2024-11-06 09:08:34.250852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.243 [2024-11-06 09:08:34.250891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.243 [2024-11-06 09:08:34.250903] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.243 [2024-11-06 09:08:34.251144] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.243 [2024-11-06 09:08:34.251366] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.243 [2024-11-06 09:08:34.251375] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.243 [2024-11-06 09:08:34.251388] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.243 [2024-11-06 09:08:34.254919] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.243 [2024-11-06 09:08:34.264051] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.243 [2024-11-06 09:08:34.264742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.243 [2024-11-06 09:08:34.264786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.243 [2024-11-06 09:08:34.264797] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.243 [2024-11-06 09:08:34.265034] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.243 [2024-11-06 09:08:34.265255] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.243 [2024-11-06 09:08:34.265265] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.243 [2024-11-06 09:08:34.265272] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.243 [2024-11-06 09:08:34.268795] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.243 [2024-11-06 09:08:34.277925] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.243 [2024-11-06 09:08:34.278518] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.243 [2024-11-06 09:08:34.278537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.243 [2024-11-06 09:08:34.278545] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.243 [2024-11-06 09:08:34.278768] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.243 [2024-11-06 09:08:34.278987] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.243 [2024-11-06 09:08:34.278996] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.243 [2024-11-06 09:08:34.279003] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.243 [2024-11-06 09:08:34.282517] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.243 [2024-11-06 09:08:34.291863] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.243 [2024-11-06 09:08:34.292551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.243 [2024-11-06 09:08:34.292590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.243 [2024-11-06 09:08:34.292601] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.243 [2024-11-06 09:08:34.292847] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.243 [2024-11-06 09:08:34.293069] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.243 [2024-11-06 09:08:34.293078] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.243 [2024-11-06 09:08:34.293086] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.243 [2024-11-06 09:08:34.296607] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.243 [2024-11-06 09:08:34.305766] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.243 [2024-11-06 09:08:34.306418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.243 [2024-11-06 09:08:34.306456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.243 [2024-11-06 09:08:34.306467] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.243 [2024-11-06 09:08:34.306704] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.243 [2024-11-06 09:08:34.306934] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.243 [2024-11-06 09:08:34.306944] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.243 [2024-11-06 09:08:34.306951] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.243 [2024-11-06 09:08:34.310475] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.243 [2024-11-06 09:08:34.319605] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.243 [2024-11-06 09:08:34.320270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.243 [2024-11-06 09:08:34.320309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.243 [2024-11-06 09:08:34.320320] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.243 [2024-11-06 09:08:34.320557] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.243 [2024-11-06 09:08:34.320787] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.243 [2024-11-06 09:08:34.320797] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.243 [2024-11-06 09:08:34.320805] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.243 [2024-11-06 09:08:34.324324] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.243 [2024-11-06 09:08:34.333455] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.243 [2024-11-06 09:08:34.334005] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.243 [2024-11-06 09:08:34.334025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.243 [2024-11-06 09:08:34.334033] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.243 [2024-11-06 09:08:34.334251] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.243 [2024-11-06 09:08:34.334469] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.243 [2024-11-06 09:08:34.334478] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.243 [2024-11-06 09:08:34.334485] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.243 [2024-11-06 09:08:34.338006] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.519 [2024-11-06 09:08:34.347336] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.519 [2024-11-06 09:08:34.347878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.519 [2024-11-06 09:08:34.347895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.519 [2024-11-06 09:08:34.347907] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.519 [2024-11-06 09:08:34.348125] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.519 [2024-11-06 09:08:34.348342] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.519 [2024-11-06 09:08:34.348350] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.519 [2024-11-06 09:08:34.348357] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.519 [2024-11-06 09:08:34.351875] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.519 [2024-11-06 09:08:34.361200] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.519 [2024-11-06 09:08:34.361794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.519 [2024-11-06 09:08:34.361818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.519 [2024-11-06 09:08:34.361827] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.519 [2024-11-06 09:08:34.362049] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.519 [2024-11-06 09:08:34.362268] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.519 [2024-11-06 09:08:34.362277] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.519 [2024-11-06 09:08:34.362283] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.519 [2024-11-06 09:08:34.365808] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.519 [2024-11-06 09:08:34.375138] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.519 [2024-11-06 09:08:34.375809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.519 [2024-11-06 09:08:34.375847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.519 [2024-11-06 09:08:34.375860] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.519 [2024-11-06 09:08:34.376099] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.519 [2024-11-06 09:08:34.376321] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.519 [2024-11-06 09:08:34.376330] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.519 [2024-11-06 09:08:34.376338] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.519 [2024-11-06 09:08:34.379867] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.519 [2024-11-06 09:08:34.389000] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.519 [2024-11-06 09:08:34.389569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.519 [2024-11-06 09:08:34.389589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.519 [2024-11-06 09:08:34.389596] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.519 [2024-11-06 09:08:34.389819] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.519 [2024-11-06 09:08:34.390043] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.519 [2024-11-06 09:08:34.390051] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.520 [2024-11-06 09:08:34.390058] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.520 [2024-11-06 09:08:34.393586] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.520 [2024-11-06 09:08:34.402942] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.520 [2024-11-06 09:08:34.403509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.520 [2024-11-06 09:08:34.403525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.520 [2024-11-06 09:08:34.403533] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.520 [2024-11-06 09:08:34.403756] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.520 [2024-11-06 09:08:34.403975] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.520 [2024-11-06 09:08:34.403984] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.520 [2024-11-06 09:08:34.403991] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.520 [2024-11-06 09:08:34.407509] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.520 [2024-11-06 09:08:34.416841] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.520 [2024-11-06 09:08:34.417455] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.520 [2024-11-06 09:08:34.417493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.520 [2024-11-06 09:08:34.417504] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.520 [2024-11-06 09:08:34.417741] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.520 [2024-11-06 09:08:34.417971] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.520 [2024-11-06 09:08:34.417981] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.520 [2024-11-06 09:08:34.417989] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.520 [2024-11-06 09:08:34.421509] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.520 [2024-11-06 09:08:34.430639] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.520 [2024-11-06 09:08:34.431194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.520 [2024-11-06 09:08:34.431213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.520 [2024-11-06 09:08:34.431221] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.520 [2024-11-06 09:08:34.431439] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.520 [2024-11-06 09:08:34.431657] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.520 [2024-11-06 09:08:34.431665] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.520 [2024-11-06 09:08:34.431677] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.520 [2024-11-06 09:08:34.435201] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.520 [2024-11-06 09:08:34.444532] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.520 [2024-11-06 09:08:34.445097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.520 [2024-11-06 09:08:34.445136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.520 [2024-11-06 09:08:34.445148] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.520 [2024-11-06 09:08:34.445384] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.520 [2024-11-06 09:08:34.445606] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.520 [2024-11-06 09:08:34.445615] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.520 [2024-11-06 09:08:34.445622] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.520 [2024-11-06 09:08:34.449150] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.520 [2024-11-06 09:08:34.458485] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.520 [2024-11-06 09:08:34.459132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.520 [2024-11-06 09:08:34.459171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.520 [2024-11-06 09:08:34.459182] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.520 [2024-11-06 09:08:34.459419] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.520 [2024-11-06 09:08:34.459641] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.520 [2024-11-06 09:08:34.459650] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.520 [2024-11-06 09:08:34.459658] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.520 [2024-11-06 09:08:34.463185] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.520 [2024-11-06 09:08:34.472310] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.520 [2024-11-06 09:08:34.472950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.520 [2024-11-06 09:08:34.472989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.520 [2024-11-06 09:08:34.473001] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.520 [2024-11-06 09:08:34.473242] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.520 [2024-11-06 09:08:34.473464] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.520 [2024-11-06 09:08:34.473472] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.520 [2024-11-06 09:08:34.473481] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.520 [2024-11-06 09:08:34.477012] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.520 [2024-11-06 09:08:34.486157] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.520 [2024-11-06 09:08:34.486764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.520 [2024-11-06 09:08:34.486783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.520 [2024-11-06 09:08:34.486791] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.520 [2024-11-06 09:08:34.487009] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.520 [2024-11-06 09:08:34.487226] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.520 [2024-11-06 09:08:34.487235] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.520 [2024-11-06 09:08:34.487242] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.520 [2024-11-06 09:08:34.490772] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.520 [2024-11-06 09:08:34.500105] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.520 [2024-11-06 09:08:34.500559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.520 [2024-11-06 09:08:34.500575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.520 [2024-11-06 09:08:34.500583] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.520 [2024-11-06 09:08:34.500806] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.520 [2024-11-06 09:08:34.501024] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.520 [2024-11-06 09:08:34.501033] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.520 [2024-11-06 09:08:34.501040] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.520 [2024-11-06 09:08:34.504562] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.520 [2024-11-06 09:08:34.513898] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.520 [2024-11-06 09:08:34.514443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.520 [2024-11-06 09:08:34.514482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.520 [2024-11-06 09:08:34.514492] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.520 [2024-11-06 09:08:34.514730] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.520 [2024-11-06 09:08:34.514960] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.520 [2024-11-06 09:08:34.514970] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.520 [2024-11-06 09:08:34.514978] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.520 [2024-11-06 09:08:34.518497] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.520 [2024-11-06 09:08:34.527836] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.520 [2024-11-06 09:08:34.528484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.520 [2024-11-06 09:08:34.528522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.520 [2024-11-06 09:08:34.528538] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.520 [2024-11-06 09:08:34.528782] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.520 [2024-11-06 09:08:34.529005] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.520 [2024-11-06 09:08:34.529014] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.521 [2024-11-06 09:08:34.529022] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.521 [2024-11-06 09:08:34.532540] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.521 [2024-11-06 09:08:34.541668] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.521 [2024-11-06 09:08:34.542319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.521 [2024-11-06 09:08:34.542358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.521 [2024-11-06 09:08:34.542369] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.521 [2024-11-06 09:08:34.542606] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.521 [2024-11-06 09:08:34.542836] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.521 [2024-11-06 09:08:34.542847] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.521 [2024-11-06 09:08:34.542855] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.521 [2024-11-06 09:08:34.546378] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.521 [2024-11-06 09:08:34.555509] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.521 [2024-11-06 09:08:34.556131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.521 [2024-11-06 09:08:34.556169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.521 [2024-11-06 09:08:34.556181] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.521 [2024-11-06 09:08:34.556419] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.521 [2024-11-06 09:08:34.556641] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.521 [2024-11-06 09:08:34.556650] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.521 [2024-11-06 09:08:34.556659] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.521 [2024-11-06 09:08:34.560190] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.521 [2024-11-06 09:08:34.569321] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.521 [2024-11-06 09:08:34.570077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.521 [2024-11-06 09:08:34.570116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.521 [2024-11-06 09:08:34.570127] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.521 [2024-11-06 09:08:34.570365] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.521 [2024-11-06 09:08:34.570592] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.521 [2024-11-06 09:08:34.570601] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.521 [2024-11-06 09:08:34.570610] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.521 [2024-11-06 09:08:34.574139] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.521 [2024-11-06 09:08:34.583266] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.521 [2024-11-06 09:08:34.583867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.521 [2024-11-06 09:08:34.583906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.521 [2024-11-06 09:08:34.583917] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.521 [2024-11-06 09:08:34.584154] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.521 [2024-11-06 09:08:34.584376] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.521 [2024-11-06 09:08:34.584384] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.521 [2024-11-06 09:08:34.584392] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.521 [2024-11-06 09:08:34.587921] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.521 [2024-11-06 09:08:34.597059] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.521 [2024-11-06 09:08:34.597764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.521 [2024-11-06 09:08:34.597802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.521 [2024-11-06 09:08:34.597814] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.521 [2024-11-06 09:08:34.598054] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.521 [2024-11-06 09:08:34.598275] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.521 [2024-11-06 09:08:34.598284] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.521 [2024-11-06 09:08:34.598292] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.521 [2024-11-06 09:08:34.601819] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.521 [2024-11-06 09:08:34.610963] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.521 [2024-11-06 09:08:34.611526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.521 [2024-11-06 09:08:34.611564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.521 [2024-11-06 09:08:34.611577] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.521 [2024-11-06 09:08:34.611822] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.521 [2024-11-06 09:08:34.612044] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.521 [2024-11-06 09:08:34.612054] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.521 [2024-11-06 09:08:34.612062] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.521 [2024-11-06 09:08:34.615588] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.521 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:32:44.521 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@864 -- # return 0 00:32:44.521 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:32:44.521 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@730 -- # xtrace_disable 00:32:44.521 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:44.521 [2024-11-06 09:08:34.624930] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.785 [2024-11-06 09:08:34.625376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.785 [2024-11-06 09:08:34.625396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.785 [2024-11-06 09:08:34.625404] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.785 [2024-11-06 09:08:34.625622] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.785 [2024-11-06 09:08:34.625846] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.785 [2024-11-06 09:08:34.625855] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.785 [2024-11-06 09:08:34.625862] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.785 [2024-11-06 09:08:34.629384] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.785 [2024-11-06 09:08:34.638729] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.785 [2024-11-06 09:08:34.639177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.785 [2024-11-06 09:08:34.639196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.785 [2024-11-06 09:08:34.639204] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.785 [2024-11-06 09:08:34.639423] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.785 [2024-11-06 09:08:34.639641] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.785 [2024-11-06 09:08:34.639650] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.785 [2024-11-06 09:08:34.639657] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.785 [2024-11-06 09:08:34.643182] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.785 [2024-11-06 09:08:34.652519] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.785 [2024-11-06 09:08:34.653039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.785 [2024-11-06 09:08:34.653055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.785 [2024-11-06 09:08:34.653063] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.785 [2024-11-06 09:08:34.653281] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.785 [2024-11-06 09:08:34.653499] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.785 [2024-11-06 09:08:34.653507] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.785 [2024-11-06 09:08:34.653518] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.785 [2024-11-06 09:08:34.657043] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:44.785 [2024-11-06 09:08:34.664195] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:32:44.785 [2024-11-06 09:08:34.666382] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.785 [2024-11-06 09:08:34.666858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.785 [2024-11-06 09:08:34.666896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.785 [2024-11-06 09:08:34.666908] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.785 [2024-11-06 09:08:34.667149] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.785 [2024-11-06 09:08:34.667370] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.785 [2024-11-06 09:08:34.667379] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.785 [2024-11-06 09:08:34.667387] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:44.785 [2024-11-06 09:08:34.670915] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.785 [2024-11-06 09:08:34.680278] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.785 [2024-11-06 09:08:34.680860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.785 [2024-11-06 09:08:34.680898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.785 [2024-11-06 09:08:34.680911] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.785 [2024-11-06 09:08:34.681149] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.785 [2024-11-06 09:08:34.681370] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.785 [2024-11-06 09:08:34.681380] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.785 [2024-11-06 09:08:34.681388] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.785 [2024-11-06 09:08:34.684917] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.785 [2024-11-06 09:08:34.694057] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.785 [2024-11-06 09:08:34.694578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.785 [2024-11-06 09:08:34.694619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.785 [2024-11-06 09:08:34.694632] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.785 [2024-11-06 09:08:34.694882] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.785 [2024-11-06 09:08:34.695104] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.785 [2024-11-06 09:08:34.695112] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.785 [2024-11-06 09:08:34.695120] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.785 Malloc0 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:44.785 [2024-11-06 09:08:34.698641] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:44.785 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:44.785 [2024-11-06 09:08:34.707990] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.785 [2024-11-06 09:08:34.708568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.785 [2024-11-06 09:08:34.708605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.785 [2024-11-06 09:08:34.708617] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.785 [2024-11-06 09:08:34.708861] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.785 [2024-11-06 09:08:34.709083] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.785 [2024-11-06 09:08:34.709092] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.786 [2024-11-06 09:08:34.709100] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.786 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:44.786 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:32:44.786 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:44.786 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:44.786 [2024-11-06 09:08:34.712621] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.786 [2024-11-06 09:08:34.721752] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.786 [2024-11-06 09:08:34.722328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:32:44.786 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:44.786 [2024-11-06 09:08:34.722347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0xe24000 with addr=10.0.0.2, port=4420 00:32:44.786 [2024-11-06 09:08:34.722355] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xe24000 is same with the state(6) to be set 00:32:44.786 [2024-11-06 09:08:34.722573] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xe24000 (9): Bad file descriptor 00:32:44.786 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:32:44.786 [2024-11-06 09:08:34.722797] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Ctrlr is in error state 00:32:44.786 [2024-11-06 09:08:34.722811] nvme_ctrlr.c:1799:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] controller reinitialization failed 00:32:44.786 [2024-11-06 09:08:34.722818] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] in failed state. 00:32:44.786 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:44.786 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:44.786 [2024-11-06 09:08:34.726333] bdev_nvme.c:2234:_bdev_nvme_reset_ctrlr_complete: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller failed. 00:32:44.786 [2024-11-06 09:08:34.729500] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:32:44.786 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:44.786 09:08:34 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@38 -- # wait 2072948 00:32:44.786 [2024-11-06 09:08:34.735670] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] resetting controller 00:32:44.786 [2024-11-06 09:08:34.769289] bdev_nvme.c:2236:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] Resetting controller successful. 00:32:46.299 4762.29 IOPS, 18.60 MiB/s [2024-11-06T08:08:37.353Z] 5599.38 IOPS, 21.87 MiB/s [2024-11-06T08:08:38.296Z] 6208.89 IOPS, 24.25 MiB/s [2024-11-06T08:08:39.239Z] 6684.10 IOPS, 26.11 MiB/s [2024-11-06T08:08:40.623Z] 7097.00 IOPS, 27.72 MiB/s [2024-11-06T08:08:41.566Z] 7426.83 IOPS, 29.01 MiB/s [2024-11-06T08:08:42.508Z] 7709.85 IOPS, 30.12 MiB/s [2024-11-06T08:08:43.449Z] 8015.43 IOPS, 31.31 MiB/s 00:32:53.336 Latency(us) 00:32:53.336 [2024-11-06T08:08:43.450Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:32:53.337 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:32:53.337 Verification LBA range: start 0x0 length 0x4000 00:32:53.337 Nvme1n1 : 15.01 8223.18 32.12 9845.42 0.00 7059.00 781.65 15510.19 00:32:53.337 [2024-11-06T08:08:43.450Z] =================================================================================================================== 00:32:53.337 [2024-11-06T08:08:43.450Z] Total : 8223.18 32.12 9845.42 0.00 7059.00 781.65 15510.19 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@39 -- # sync 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@561 -- # xtrace_disable 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@42 -- # trap - SIGINT SIGTERM EXIT 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- host/bdevperf.sh@44 -- # nvmftestfini 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@514 -- # nvmfcleanup 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@121 -- # sync 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@124 -- # set +e 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@125 -- # for i in {1..20} 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:32:53.337 rmmod nvme_tcp 00:32:53.337 rmmod nvme_fabrics 00:32:53.337 rmmod nvme_keyring 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@128 -- # set -e 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@129 -- # return 0 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@515 -- # '[' -n 2073990 ']' 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@516 -- # killprocess 2073990 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@950 -- # '[' -z 2073990 ']' 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@954 -- # kill -0 2073990 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@955 -- # uname 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:32:53.337 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2073990 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2073990' 00:32:53.598 killing process with pid 2073990 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@969 -- # kill 2073990 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@974 -- # wait 2073990 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@297 -- # iptr 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@789 -- # iptables-restore 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@789 -- # iptables-save 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@302 -- # remove_spdk_ns 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:53.598 09:08:43 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:56.145 09:08:45 nvmf_tcp.nvmf_host.nvmf_bdevperf -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:32:56.145 00:32:56.145 real 0m27.889s 00:32:56.145 user 1m3.368s 00:32:56.145 sys 0m7.187s 00:32:56.145 09:08:45 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@1126 -- # xtrace_disable 00:32:56.145 09:08:45 nvmf_tcp.nvmf_host.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:32:56.145 ************************************ 00:32:56.145 END TEST nvmf_bdevperf 00:32:56.145 ************************************ 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@48 -- # run_test nvmf_target_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1107 -- # xtrace_disable 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:32:56.146 ************************************ 00:32:56.146 START TEST nvmf_target_disconnect 00:32:56.146 ************************************ 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:32:56.146 * Looking for test storage... 00:32:56.146 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1689 -- # lcov --version 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@333 -- # local ver1 ver1_l 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@334 -- # local ver2 ver2_l 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@336 -- # IFS=.-: 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@336 -- # read -ra ver1 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@337 -- # IFS=.-: 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@337 -- # read -ra ver2 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@338 -- # local 'op=<' 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@340 -- # ver1_l=2 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@341 -- # ver2_l=1 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@344 -- # case "$op" in 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@345 -- # : 1 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@364 -- # (( v = 0 )) 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@365 -- # decimal 1 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@353 -- # local d=1 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@355 -- # echo 1 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@365 -- # ver1[v]=1 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@366 -- # decimal 2 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@353 -- # local d=2 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@355 -- # echo 2 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@366 -- # ver2[v]=2 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@368 -- # return 0 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:32:56.146 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:56.146 --rc genhtml_branch_coverage=1 00:32:56.146 --rc genhtml_function_coverage=1 00:32:56.146 --rc genhtml_legend=1 00:32:56.146 --rc geninfo_all_blocks=1 00:32:56.146 --rc geninfo_unexecuted_blocks=1 00:32:56.146 00:32:56.146 ' 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:32:56.146 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:56.146 --rc genhtml_branch_coverage=1 00:32:56.146 --rc genhtml_function_coverage=1 00:32:56.146 --rc genhtml_legend=1 00:32:56.146 --rc geninfo_all_blocks=1 00:32:56.146 --rc geninfo_unexecuted_blocks=1 00:32:56.146 00:32:56.146 ' 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:32:56.146 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:56.146 --rc genhtml_branch_coverage=1 00:32:56.146 --rc genhtml_function_coverage=1 00:32:56.146 --rc genhtml_legend=1 00:32:56.146 --rc geninfo_all_blocks=1 00:32:56.146 --rc geninfo_unexecuted_blocks=1 00:32:56.146 00:32:56.146 ' 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:32:56.146 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:32:56.146 --rc genhtml_branch_coverage=1 00:32:56.146 --rc genhtml_function_coverage=1 00:32:56.146 --rc genhtml_legend=1 00:32:56.146 --rc geninfo_all_blocks=1 00:32:56.146 --rc geninfo_unexecuted_blocks=1 00:32:56.146 00:32:56.146 ' 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@7 -- # uname -s 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@15 -- # shopt -s extglob 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@5 -- # export PATH 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@51 -- # : 0 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:32:56.146 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:32:56.147 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@55 -- # have_pci_nics=0 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@11 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@13 -- # MALLOC_BDEV_SIZE=64 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@69 -- # nvmftestinit 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@474 -- # prepare_net_devs 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@436 -- # local -g is_hw=no 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@438 -- # remove_spdk_ns 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@309 -- # xtrace_disable 00:32:56.147 09:08:45 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@315 -- # pci_devs=() 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@315 -- # local -a pci_devs 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@316 -- # pci_net_devs=() 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@317 -- # pci_drivers=() 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@317 -- # local -A pci_drivers 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@319 -- # net_devs=() 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@319 -- # local -ga net_devs 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@320 -- # e810=() 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@320 -- # local -ga e810 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@321 -- # x722=() 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@321 -- # local -ga x722 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@322 -- # mlx=() 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@322 -- # local -ga mlx 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:33:04.295 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:33:04.295 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:04.295 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@416 -- # [[ up == up ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:33:04.296 Found net devices under 0000:4b:00.0: cvl_0_0 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@416 -- # [[ up == up ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:33:04.296 Found net devices under 0000:4b:00.1: cvl_0_1 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@440 -- # is_hw=yes 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:33:04.296 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:33:04.296 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.667 ms 00:33:04.296 00:33:04.296 --- 10.0.0.2 ping statistics --- 00:33:04.296 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:04.296 rtt min/avg/max/mdev = 0.667/0.667/0.667/0.000 ms 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:33:04.296 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:33:04.296 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.292 ms 00:33:04.296 00:33:04.296 --- 10.0.0.1 ping statistics --- 00:33:04.296 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:04.296 rtt min/avg/max/mdev = 0.292/0.292/0.292/0.000 ms 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@448 -- # return 0 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@70 -- # run_test nvmf_target_disconnect_tc1 nvmf_target_disconnect_tc1 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:33:04.296 ************************************ 00:33:04.296 START TEST nvmf_target_disconnect_tc1 00:33:04.296 ************************************ 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@1125 -- # nvmf_target_disconnect_tc1 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- host/target_disconnect.sh@32 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@650 -- # local es=0 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:33:04.296 [2024-11-06 09:08:53.597705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:04.296 [2024-11-06 09:08:53.597798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1a13ad0 with addr=10.0.0.2, port=4420 00:33:04.296 [2024-11-06 09:08:53.597832] nvme_tcp.c:2612:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:33:04.296 [2024-11-06 09:08:53.597844] nvme.c: 831:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:33:04.296 [2024-11-06 09:08:53.597852] nvme.c: 939:spdk_nvme_probe_ext: *ERROR*: Create probe context failed 00:33:04.296 spdk_nvme_probe() failed for transport address '10.0.0.2' 00:33:04.296 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect: errors occurred 00:33:04.296 Initializing NVMe Controllers 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@653 -- # es=1 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:33:04.296 00:33:04.296 real 0m0.128s 00:33:04.296 user 0m0.064s 00:33:04.296 sys 0m0.064s 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@10 -- # set +x 00:33:04.296 ************************************ 00:33:04.296 END TEST nvmf_target_disconnect_tc1 00:33:04.296 ************************************ 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@71 -- # run_test nvmf_target_disconnect_tc2 nvmf_target_disconnect_tc2 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:33:04.296 ************************************ 00:33:04.296 START TEST nvmf_target_disconnect_tc2 00:33:04.296 ************************************ 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@1125 -- # nvmf_target_disconnect_tc2 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@37 -- # disconnect_init 10.0.0.2 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@507 -- # nvmfpid=2080145 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@508 -- # waitforlisten 2080145 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@831 -- # '[' -z 2080145 ']' 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:04.296 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:04.296 09:08:53 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:04.296 [2024-11-06 09:08:53.770143] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:33:04.296 [2024-11-06 09:08:53.770208] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:04.296 [2024-11-06 09:08:53.869181] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:33:04.296 [2024-11-06 09:08:53.920784] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:33:04.296 [2024-11-06 09:08:53.920833] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:33:04.296 [2024-11-06 09:08:53.920842] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:33:04.296 [2024-11-06 09:08:53.920853] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:33:04.296 [2024-11-06 09:08:53.920859] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:33:04.296 [2024-11-06 09:08:53.922910] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:33:04.296 [2024-11-06 09:08:53.923140] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:33:04.296 [2024-11-06 09:08:53.923311] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:33:04.296 [2024-11-06 09:08:53.923312] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:33:04.558 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:04.558 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@864 -- # return 0 00:33:04.558 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:33:04.558 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:04.558 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:04.558 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:33:04.558 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:33:04.558 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:04.558 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:04.819 Malloc0 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:04.819 [2024-11-06 09:08:54.680379] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:04.819 [2024-11-06 09:08:54.720776] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@42 -- # reconnectpid=2080373 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:33:04.819 09:08:54 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@44 -- # sleep 2 00:33:06.738 09:08:56 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@45 -- # kill -9 2080145 00:33:06.738 09:08:56 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@47 -- # sleep 2 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 [2024-11-06 09:08:56.753807] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Write completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.738 Read completed with error (sct=0, sc=8) 00:33:06.738 starting I/O failed 00:33:06.739 Write completed with error (sct=0, sc=8) 00:33:06.739 starting I/O failed 00:33:06.739 Read completed with error (sct=0, sc=8) 00:33:06.739 starting I/O failed 00:33:06.739 [2024-11-06 09:08:56.754115] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:06.739 [2024-11-06 09:08:56.754482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.754499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.754832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.754844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.755153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.755192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.755542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.755555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.755964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.756002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.756359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.756372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.756717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.756728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.757147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.757190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.757407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.757420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.757609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.757621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.757701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.757711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.758063] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.758075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.758410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.758420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.758720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.758731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.759068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.759079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.759391] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.759402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.759587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.759597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.759680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.759690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.759993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.760005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.760328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.760339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.760634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.760644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.760837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.760847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.761153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.761164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.761451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.761461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.761739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.761762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.762191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.762201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.762495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.762506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.762848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.762859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.763054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.763066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.763258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.763269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.763444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.763454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.763514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.763524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.763762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.763774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.764086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.764096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.764373] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.764386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.764560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.764571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.764877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.739 [2024-11-06 09:08:56.764887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.739 qpair failed and we were unable to recover it. 00:33:06.739 [2024-11-06 09:08:56.765212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.765221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.765540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.765550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.765853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.765864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.766064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.766074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.766258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.766268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.766438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.766447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.766710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.766720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.767048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.767058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.767355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.767366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.767526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.767537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.767887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.767898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.768081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.768092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.768315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.768325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.768512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.768522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.768800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.768809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.769016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.769026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.769313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.769322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.769490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.769501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.769814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.769824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.770101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.770111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.770426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.770437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.770638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.770648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.770988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.770998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.771205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.771214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.771512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.771522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.771811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.771821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.772124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.772134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.772500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.772511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.772777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.772787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.773090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.773099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.773393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.773403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.773614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.773623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.773931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.773942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.774212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.774222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.774602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.774612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.774898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.774908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.775099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.775108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.775451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.775461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.740 [2024-11-06 09:08:56.775733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.740 [2024-11-06 09:08:56.775754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.740 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.775927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.775936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.776133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.776143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.776459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.776469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.776768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.776778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.777058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.777068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.777370] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.777381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.777652] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.777661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.778007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.778018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.778352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.778362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.778646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.778656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.778975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.778985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.779274] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.779284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.779563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.779572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.779896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.779906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.780202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.780211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.780505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.780514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.780798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.780808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.781102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.781112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.781424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.781433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.781595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.781606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.781781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.781791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.782141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.782151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.782440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.782450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.782733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.782743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.783063] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.783074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.783360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.783370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.783564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.783577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.783931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.783942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.784234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.784244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.784519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.784528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.784813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.784824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.785116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.785126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.785409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.785419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.785747] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.785758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.786130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.786139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.786445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.786454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.786742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.786754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.787076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.787086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.741 qpair failed and we were unable to recover it. 00:33:06.741 [2024-11-06 09:08:56.787413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.741 [2024-11-06 09:08:56.787423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.787706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.787716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.788051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.788063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.788354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.788364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.788649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.788660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.788966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.788978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.789257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.789267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.789574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.789584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.789865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.789875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.790159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.790168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.790457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.790468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.790795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.790805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.791141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.791151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.791482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.791492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.791784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.791794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.792121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.792131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.792406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.792416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.792726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.792736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.793073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.793084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.793385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.793395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.793737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.793750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.794119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.794129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.794423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.794432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.794721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.794731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.795014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.795025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.795335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.795345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.795676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.795685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.795986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.795996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.796291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.796301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.796581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.796593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.796872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.796882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.797259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.797269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.797571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.797580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.797677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.797686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.797960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.797970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.798308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.798318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.742 qpair failed and we were unable to recover it. 00:33:06.742 [2024-11-06 09:08:56.798598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.742 [2024-11-06 09:08:56.798608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.798908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.798918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.799223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.799233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.799531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.799541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.799819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.799830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.800135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.800145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.800439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.800449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.800753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.800764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.800959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.800970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.801269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.801279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.801570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.801579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.801878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.801888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.802181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.802191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.802532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.802542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.802915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.802925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.803203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.803213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.803506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.803516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.803808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.803818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.804119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.804128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.804436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.804446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.804738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.804751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.805034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.805044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.805416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.805425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.805722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.805731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.806068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.806078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.806360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.806369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.806662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.806671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.807050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.807060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.807365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.807375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.807688] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.807698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.807991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.808003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.808346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.808357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.808700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.808710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.808930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.808940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.809242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.743 [2024-11-06 09:08:56.809252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.743 qpair failed and we were unable to recover it. 00:33:06.743 [2024-11-06 09:08:56.809548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.809559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.809892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.809902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.810217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.810227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.810529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.810538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.810813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.810824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.811136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.811145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.811456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.811466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.811760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.811771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.811962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.811972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.812363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.812373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.812668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.812678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.812974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.812984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.813286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.813295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.813598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.813609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.813860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.813871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.814179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.814189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.814488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.814497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.814809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.814819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.815177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.815187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.815462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.815472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.815796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.815806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.816091] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.816101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.816315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.816325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.816691] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.816701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.816987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.816997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.817372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.817381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.817727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.817738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.818040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.818050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.818223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.818233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.818522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.818532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.818865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.818875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.819178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.819188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.819487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.819498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.819884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.819894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.820197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.820207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.820588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.820597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.820797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.820807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.821093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.744 [2024-11-06 09:08:56.821103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.744 qpair failed and we were unable to recover it. 00:33:06.744 [2024-11-06 09:08:56.821284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.821294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.821592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.821602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.821816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.821827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.822126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.822136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.822458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.822468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.822789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.822800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.823105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.823115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.823421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.823430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.823816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.823826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.824133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.824143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.824499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.824508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.824811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.824821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.825120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.825130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.825443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.825453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.825809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.825820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.826124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.826134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.826333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.826342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.826658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.826668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.827006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.827016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.827216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.827225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.827535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.827545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.827842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.827854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.828157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.828167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.828504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.828514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.828810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.828820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.829135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.829145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.829425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.829435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.829763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.829773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.830059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.830069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.830352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.830364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.830717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.830727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.831066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.831076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.831367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.831377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.831679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.831689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.831994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.832005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.832311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.832321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.832513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.832523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.832798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.832808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.833036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.745 [2024-11-06 09:08:56.833046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.745 qpair failed and we were unable to recover it. 00:33:06.745 [2024-11-06 09:08:56.833345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.833355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.833655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.833665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.833973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.833983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.834219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.834228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.834555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.834565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.834846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.834856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.835065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.835075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.835270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.835280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.835465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.835476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.835759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.835771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.835996] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.836006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.836292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.836302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.836625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.836635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.836858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.836868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.837207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.837217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.837539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.837549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.837894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.837905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.838285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.838297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.838649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.838659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.838979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.838989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.839363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.839374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.839680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.839690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.839893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.839903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.840077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.840087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.840287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.840297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.840607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.840616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.841016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.841027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.841309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.841319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.841624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.841635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.746 qpair failed and we were unable to recover it. 00:33:06.746 [2024-11-06 09:08:56.842049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.746 [2024-11-06 09:08:56.842059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.747 qpair failed and we were unable to recover it. 00:33:06.747 [2024-11-06 09:08:56.842273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.747 [2024-11-06 09:08:56.842283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.747 qpair failed and we were unable to recover it. 00:33:06.747 [2024-11-06 09:08:56.842464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.747 [2024-11-06 09:08:56.842474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.747 qpair failed and we were unable to recover it. 00:33:06.747 [2024-11-06 09:08:56.842685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.747 [2024-11-06 09:08:56.842695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.747 qpair failed and we were unable to recover it. 00:33:06.747 [2024-11-06 09:08:56.842923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.747 [2024-11-06 09:08:56.842933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.747 qpair failed and we were unable to recover it. 00:33:06.747 [2024-11-06 09:08:56.843290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.747 [2024-11-06 09:08:56.843300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.747 qpair failed and we were unable to recover it. 00:33:06.747 [2024-11-06 09:08:56.843673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.747 [2024-11-06 09:08:56.843683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.747 qpair failed and we were unable to recover it. 00:33:06.747 [2024-11-06 09:08:56.843978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:06.747 [2024-11-06 09:08:56.843988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:06.747 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.844297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.844309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.844524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.844536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.844716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.844727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.845043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.845054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.845391] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.845402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.845708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.845718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.846028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.846039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.846324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.846334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.846666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.846677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.847075] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.847086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.847415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.847426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.847615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.847626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.847915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.847925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.848214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.848225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.848523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.848534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.848817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.848827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.849187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.849197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.849479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.849489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.849793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.849803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.850110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.850120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.850428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.850437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.850733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.850749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.851089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.023 [2024-11-06 09:08:56.851099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.023 qpair failed and we were unable to recover it. 00:33:07.023 [2024-11-06 09:08:56.851402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.851412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.851766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.851777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.852078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.852088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.852395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.852405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.852580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.852592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.852893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.852904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.853178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.853189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.853488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.853498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.853808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.853818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.854106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.854116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.854435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.854445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.854754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.854764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.855131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.855142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.855448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.855459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.855792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.855803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.856082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.856092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.856394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.856403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.856693] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.856703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.856988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.856999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.857306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.857316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.857595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.857605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.857920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.857930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.858212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.858230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.858526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.858536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.858813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.858823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.859142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.859154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.859494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.859504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.859805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.859815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.860109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.860118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.860427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.860437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.860718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.860728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.861050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.861060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.861349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.861359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.861660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.861670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.861977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.861987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.862273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.862283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.024 [2024-11-06 09:08:56.862586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.024 [2024-11-06 09:08:56.862596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.024 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.862913] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.862923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.863211] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.863221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.863532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.863542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.863850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.863860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.864181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.864191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.864494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.864504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.864805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.864816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.865031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.865040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.865310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.865320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.865643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.865653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.865957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.865967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.866277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.866287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.866593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.866603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.866802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.866815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.867166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.867176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.867473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.867483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.867661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.867673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.868001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.868012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.868320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.868330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.868662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.868672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.868982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.868993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.869174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.869185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.869484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.869494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.869689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.869700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.870009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.870019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.870306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.870316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.870625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.870634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.870920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.870931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.871124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.871134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.871464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.871477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.871785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.871796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.872091] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.872100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.872405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.872415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.872732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.872743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.873045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.873056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.025 [2024-11-06 09:08:56.873337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.025 [2024-11-06 09:08:56.873354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.025 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.873676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.873686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.874022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.874033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.874299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.874309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.874615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.874625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.874914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.874924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.875237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.875247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.875532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.875542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.875859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.875869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.876192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.876203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.876521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.876531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.876843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.876853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.877067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.877077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.877377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.877387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.877694] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.877704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.878049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.878059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.878353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.878363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.878644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.878654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.878933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.878943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.879264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.879274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.879510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.879520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.879724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.879737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.880091] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.880101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.880393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.880410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.880727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.880737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.881049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.881059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.881283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.881293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.881610] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.881621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.881922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.881933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.882216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.882227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.882530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.882539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.882854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.882865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.883173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.883182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.883485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.883495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.883842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.883853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.884205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.884215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.026 [2024-11-06 09:08:56.884519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.026 [2024-11-06 09:08:56.884529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.026 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.884832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.884842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.885136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.885147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.885483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.885493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.885786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.885796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.886082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.886092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.886395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.886405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.886599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.886609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.886954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.886964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.887284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.887294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.887602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.887612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.887847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.887858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.888172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.888182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.888489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.888499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.888795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.888806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.889111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.889121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.889423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.889433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.889728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.889738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.890029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.890039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.890347] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.890357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.890688] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.890698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.891078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.891089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.891306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.891316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.891663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.891672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.891975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.891985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.892289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.892299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.892595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.892607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.892900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.892910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.893230] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.893240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.893527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.893537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.893854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.893864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.894157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.894167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.894457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.894466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.894754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.894764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.895088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.895099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.895383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.895393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.895694] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.027 [2024-11-06 09:08:56.895705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.027 qpair failed and we were unable to recover it. 00:33:07.027 [2024-11-06 09:08:56.896092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.896103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.896404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.896415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.896597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.896609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.896956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.896966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.897288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.897298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.897602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.897612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.897663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.897673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.898000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.898010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.898344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.898354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.898623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.898633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.898950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.898960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.899269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.899285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.899597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.899606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.899887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.899898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.900241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.900250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.900537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.900553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.900876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.900886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.901186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.901196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.901486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.901496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.901806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.901816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.902206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.902217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.902492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.902502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.902851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.902862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.903223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.903233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.903525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.903535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.903840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.903851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.904166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.904176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.904488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.904497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.904789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.904799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.028 [2024-11-06 09:08:56.905099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.028 [2024-11-06 09:08:56.905108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.028 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.905463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.905473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.905692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.905702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.906016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.906027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.906229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.906239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.906583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.906593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.906879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.906889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.907198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.907208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.907553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.907563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.907911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.907922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.908113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.908124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.908438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.908448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.908815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.908825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.909129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.909147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.909463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.909472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.909812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.909823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.910004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.910015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.910218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.910227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.910584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.910593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.910915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.910925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.911238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.911248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.911535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.911544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.911849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.911860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.912057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.912067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.912424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.912434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.912732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.912741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.913062] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.913072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.913355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.913365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.913561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.913578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.913909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.913920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.914206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.914216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.914497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.914507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.914776] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.914786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.915113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.915123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.915507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.915517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.915800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.915811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.916113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.916123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.029 qpair failed and we were unable to recover it. 00:33:07.029 [2024-11-06 09:08:56.916440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.029 [2024-11-06 09:08:56.916450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.916731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.916749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.917065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.917075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.917380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.917390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.917700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.917710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.918011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.918021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.918330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.918340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.918643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.918653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.918928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.918938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.919247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.919256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.919584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.919594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.919915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.919926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.920234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.920245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.920550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.920561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.920752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.920762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.921042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.921051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.921341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.921352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.921558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.921569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.921890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.921900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.922215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.922226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.922511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.922521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.922816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.922827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.923145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.923155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.923482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.923492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.923799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.923810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.924118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.924128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.924426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.924438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.924768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.924779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.925104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.925115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.925384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.925394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.925618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.925628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.925924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.925934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.926237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.926248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.926559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.926570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.926876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.926887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.927222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.927233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.927500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.927512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.927816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.927827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.030 qpair failed and we were unable to recover it. 00:33:07.030 [2024-11-06 09:08:56.928033] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.030 [2024-11-06 09:08:56.928043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.928407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.928416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.928699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.928710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.929004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.929014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.929324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.929333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.929661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.929670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.929967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.929983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.930296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.930306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.930629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.930639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.930933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.930943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.931216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.931226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.931432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.931442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.931742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.931756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.932042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.932052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.932368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.932378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.932708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.932718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.933053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.933064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.933239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.933248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.933543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.933553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.933867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.933877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.934197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.934207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.934534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.934546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.934864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.934874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.935197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.935207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.935583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.935593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.935930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.935941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.936272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.936281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.936561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.936570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.936877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.936888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.937208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.937218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.937509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.937519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.937827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.937838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.938146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.938157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.938358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.938368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.938677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.938687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.031 qpair failed and we were unable to recover it. 00:33:07.031 [2024-11-06 09:08:56.939006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.031 [2024-11-06 09:08:56.939017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.939357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.939367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.939668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.939678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.939993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.940004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.940397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.940408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.940734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.940743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.941060] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.941070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.941384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.941395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.941721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.941732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.942031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.942042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.942322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.942332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.942634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.942644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.942967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.942978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.943311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.943322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.943653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.943664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.943992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.944002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.944283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.944293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.944604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.944613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.944940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.944957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.945296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.945307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.945611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.945621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.945922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.945932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.946231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.946241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.946553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.946564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.946866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.946876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.947178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.947188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.947500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.947510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.947819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.947832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.948127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.948137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.948452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.948462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.948771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.948783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.949095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.949105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.949399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.949415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.949725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.949735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.950044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.950055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.032 [2024-11-06 09:08:56.950376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.032 [2024-11-06 09:08:56.950387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.032 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.950696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.950706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.951023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.951033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.951343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.951354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.951660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.951670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.952003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.952013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.952303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.952314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.952619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.952628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.952920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.952930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.953245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.953254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.953533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.953543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.953764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.953775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.954085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.954095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.954405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.954415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.954776] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.954787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.955105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.955115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.955426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.955436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.955721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.955732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.956064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.956074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.956467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.956479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.956698] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.956707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.957040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.957050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.957372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.957382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.957755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.957765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.957990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.958000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.958202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.958212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.958494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.958504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.958818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.958828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.958941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.958951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.959130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.959140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.959439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.959449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.959738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.959750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.960049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.960059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.960399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.960410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.960714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.960724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.033 [2024-11-06 09:08:56.961019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.033 [2024-11-06 09:08:56.961037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.033 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.961323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.961333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.961643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.961653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.961816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.961828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.962182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.962192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.962486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.962497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.962655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.962666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.962981] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.962992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.963374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.963385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.963679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.963689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.964019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.964030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.964298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.964308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.964610] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.964621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.965002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.965012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.965305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.965315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.965624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.965634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.965963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.965974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.966282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.966293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.966607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.966654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.966837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.966849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.967174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.967184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.967489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.967499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.967700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.967710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.968066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.968076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.968403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.968413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.968609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.968622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.968893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.968904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.969221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.969230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.969418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.969430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.969805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.969817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.970177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.970187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.970494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.970503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.970813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.970824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.971142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.971152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.971370] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.971380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.971698] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.971708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.972039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.972049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.972331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.034 [2024-11-06 09:08:56.972343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.034 qpair failed and we were unable to recover it. 00:33:07.034 [2024-11-06 09:08:56.972649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.972659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.972981] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.972992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.973324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.973334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.973643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.973654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.973990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.974001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.974177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.974188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.974518] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.974528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.974835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.974845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.975130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.975140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.975442] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.975452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.975758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.975768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.976084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.976094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.976397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.976407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.976727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.976738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.977080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.977093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.977411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.977421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.977740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.977758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.978044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.978054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.978247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.978257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.978572] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.978582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.978954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.978965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.979294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.979303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.979606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.979616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.979895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.979905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.980227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.980236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.980397] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.980408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.980735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.980749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.981100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.981110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.981431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.981441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.981774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.981785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.982162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.982172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.982501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.982511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.982850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.982860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.983174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.983184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.983477] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.983487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.983777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.983787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.984098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.035 [2024-11-06 09:08:56.984107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.035 qpair failed and we were unable to recover it. 00:33:07.035 [2024-11-06 09:08:56.984392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.984408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.984700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.984710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.985043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.985055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.985273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.985283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.985616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.985625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.985917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.985927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.986242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.986251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.986538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.986548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.986819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.986830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.987115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.987126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.987417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.987428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.987730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.987740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.988122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.988132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.988470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.988480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.988788] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.988798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.989110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.989120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.989488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.989498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.989806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.989816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.990123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.990136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.990466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.990481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.990807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.990817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.991121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.991131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.991455] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.991465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.991773] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.991783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.992099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.992109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.992476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.992486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.992793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.992803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.993117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.993127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.993404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.993415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.993751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.993762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.994144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.994155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.994466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.994476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.994660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.994670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.995017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.995028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.036 [2024-11-06 09:08:56.995318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.036 [2024-11-06 09:08:56.995329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.036 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.995632] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.995642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.995922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.995932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.996294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.996304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.996622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.996633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.996854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.996864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.997199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.997209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.997513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.997523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.997817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.997827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.997991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.998002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.998317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.998327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.998622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.998632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.998924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.998934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.999248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.999258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.999566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.999576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:56.999915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:56.999926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.000237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.000247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.000558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.000568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.000877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.000888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.001207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.001217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.001517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.001527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.001807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.001817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.002130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.002140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.002443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.002453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.002760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.002771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.002961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.002980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.003337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.003347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.003656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.003667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.003979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.003989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.004291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.004303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.004590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.004600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.004921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.004931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.005251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.005262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.005562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.037 [2024-11-06 09:08:57.005573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.037 qpair failed and we were unable to recover it. 00:33:07.037 [2024-11-06 09:08:57.005894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.005905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.006227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.006237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.006533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.006544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.006841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.006852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.007167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.007178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.007517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.007527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.007791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.007802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.008182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.008192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.008523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.008533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.008846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.008857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.009142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.009152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.009346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.009357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.009577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.009587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.009891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.009902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.010099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.010109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.010423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.010434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.010730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.010740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.011044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.011054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.011334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.011352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.011663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.011673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.011913] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.011924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.012286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.012295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.012528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.012537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.012764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.012775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.013097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.013107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.013415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.013424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.013711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.013721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.013939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.013950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.014171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.014181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.014525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.014535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.014734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.014744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.038 [2024-11-06 09:08:57.015006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.038 [2024-11-06 09:08:57.015016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.038 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.015365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.015375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.015593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.015603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.015888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.015898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.016118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.016128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.016372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.016382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.016725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.016736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.016952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.016962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.017364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.017374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.017580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.017591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.017900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.017911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.018100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.018110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.018344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.018354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.018662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.018672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.018947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.018957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.019178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.019188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.019498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.019507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.019842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.019853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.020177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.020188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.020513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.020522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.020732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.020742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.021076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.021086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.021414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.021424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.021531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.021541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.021767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.021777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.022080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.022090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.022398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.022407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.022756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.022767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.023104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.023114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.023406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.023417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.023616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.023626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.023917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.023928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.024233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.024243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.024546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.024556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.024865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.024875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.025189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.025200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.025347] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.025358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.025659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.025669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.039 qpair failed and we were unable to recover it. 00:33:07.039 [2024-11-06 09:08:57.026018] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.039 [2024-11-06 09:08:57.026028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.026455] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.026466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.026682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.026692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.026888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.026898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.027183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.027193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.027495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.027505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.027716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.027726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.027920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.027932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.028213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.028223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.028543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.028554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.028901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.028912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.029250] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.029261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.029595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.029605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.029893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.029903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.030226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.030236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.030550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.030560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.030767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.030777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.030996] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.031008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.031394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.031405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.031729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.031740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.032062] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.032072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.032367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.032377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.032565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.032584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.032784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.032794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.033116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.033127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.033421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.033431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.033640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.033650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.033965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.033976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.034179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.034189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.034519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.034528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.034866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.034876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.035258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.035268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.035533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.035543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.035819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.035831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.036210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.036220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.036421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.036431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.036739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.036760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.040 qpair failed and we were unable to recover it. 00:33:07.040 [2024-11-06 09:08:57.036930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.040 [2024-11-06 09:08:57.036940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.037264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.037274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.037607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.037617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.037929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.037939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.038145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.038155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.038365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.038374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.038675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.038685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.038998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.039008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.039202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.039212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.039442] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.039451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.039761] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.039771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.040087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.040097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.040270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.040282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.040617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.040626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.040964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.040975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.041288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.041299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.041646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.041656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.041973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.041984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.042279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.042290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.042468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.042477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.042778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.042789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.043151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.043166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.043487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.043497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.043700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.043710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.044014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.044025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.044089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.044100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.044354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.044365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.044658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.044669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.045035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.045046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.045300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.045310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.045648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.045659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.045981] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.045992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.046056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.046064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.046168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.046177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.046555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.046566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.046863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.046874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.041 [2024-11-06 09:08:57.047055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.041 [2024-11-06 09:08:57.047065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.041 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.047434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.047444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.047752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.047762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.048072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.048082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.048282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.048292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.048604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.048615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.048915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.048925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.049235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.049245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.049550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.049560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.049844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.049854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.050139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.050150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.050532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.050542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.050843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.050857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.051210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.051221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.051627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.051638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.051941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.051952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.052153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.052163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.052356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.052365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.052675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.052684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.052879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.052889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.053241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.053251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.053441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.053451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.053615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.053626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.053943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.053953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.054123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.054133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.054461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.054472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.054792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.054803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.055118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.055128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.055407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.055418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.055703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.055713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.056004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.056014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.056188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.056199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.042 [2024-11-06 09:08:57.056481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.042 [2024-11-06 09:08:57.056491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.042 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.056804] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.056814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.056995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.057006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.057340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.057350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.057640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.057656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.057884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.057894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.058225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.058235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.058569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.058579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.058970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.058980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.059295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.059306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.059611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.059621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.059963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.059973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.060293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.060303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.060613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.060623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.060915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.060926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.061248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.061258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.061571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.061580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.061773] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.061785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.062078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.062088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.062365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.062374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.062715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.062725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.063046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.063059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.063330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.063340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.063545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.063555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.063808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.063819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.064133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.064143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.064430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.064440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.064753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.064763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.065073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.065083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.065393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.065402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.065588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.065599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.065931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.065942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.066356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.066367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.066684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.066694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.066989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.066999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.067238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.067248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.067598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.067608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.067892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.067903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.043 qpair failed and we were unable to recover it. 00:33:07.043 [2024-11-06 09:08:57.068196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.043 [2024-11-06 09:08:57.068206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.068532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.068542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.068859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.068869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.069186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.069196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.069557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.069567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.069891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.069902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.070194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.070204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.070555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.070565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.070838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.070848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.071109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.071119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.071504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.071516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.071824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.071835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.072132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.072141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.072458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.072468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.072772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.072782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.073079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.073090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.073385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.073395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.073730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.073740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.073948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.073959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.074312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.074322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.074649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.074659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.074979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.074993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.075293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.075303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.075598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.075609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.075961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.075971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.076275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.076285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.076597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.076607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.076791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.076801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.077191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.077201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.077492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.077501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.077703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.077713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.078103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.078113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.078416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.078426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.078736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.078760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.078939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.078949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.079261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.044 [2024-11-06 09:08:57.079271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.044 qpair failed and we were unable to recover it. 00:33:07.044 [2024-11-06 09:08:57.079563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.079572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.079896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.079907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.080140] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.080150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.080342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.080353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.080622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.080632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.080953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.080963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.081168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.081177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.081494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.081505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.081804] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.081814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.082042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.082052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.082416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.082426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.082742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.082756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.082959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.082969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.083267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.083276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.083448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.083459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.083603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.083617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.083863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.083874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.084158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.084168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.084323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.084334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.084613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.084622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.084843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.084853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.085044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.085054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.085395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.085405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.085704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.085715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.086035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.086046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.086369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.086378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.086676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.086687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.086894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.086905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.087282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.087292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.087576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.087587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.087889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.087900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.088229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.088238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.088530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.088540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.088848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.088858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.089148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.089157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.089513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.089524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.089729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.089739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.090063] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.045 [2024-11-06 09:08:57.090074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.045 qpair failed and we were unable to recover it. 00:33:07.045 [2024-11-06 09:08:57.090258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.090268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.090557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.090567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.090737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.090752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.091054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.091064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.091382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.091394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.091719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.091729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.092142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.092152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.092355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.092365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.092595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.092606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.092825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.092836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.093043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.093053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.093370] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.093380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.093716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.093725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.093941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.093952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.094156] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.094166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.094291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.094301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.094589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.094599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.094892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.094902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.095213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.095222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.095533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.095543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.095718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.095728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.096055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.096066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.096426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.096436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.096719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.096729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.097072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.097082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.097371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.097382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.097721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.097730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.097952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.097963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.098305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.098314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.098585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.098596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.098813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.098824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.099015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.099025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.099303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.099312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.099654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.099665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.099941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.099951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.100260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.100270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.100584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.100595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.100906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.100916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.046 qpair failed and we were unable to recover it. 00:33:07.046 [2024-11-06 09:08:57.101238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.046 [2024-11-06 09:08:57.101248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.101573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.101583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.101890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.101900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.102203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.102212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.102532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.102541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.102856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.102867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.103180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.103190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.103528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.103540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.103822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.103833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.104162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.104172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.104452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.104469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.104805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.104816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.105105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.105115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.105422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.105431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.105715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.105725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.106027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.106037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.106322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.106332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.106651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.106661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.106967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.106979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.107265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.107275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.107582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.107592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.107896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.107907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.108115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.108125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.108305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.108317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.108526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.108536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.108823] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.108833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.109045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.109056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.109395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.109406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.109579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.109589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.109920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.047 [2024-11-06 09:08:57.109931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.047 qpair failed and we were unable to recover it. 00:33:07.047 [2024-11-06 09:08:57.110133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.110144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.110448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.110458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.110755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.110766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.111079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.111090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.111390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.111401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.111737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.111757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.112064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.112074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.112390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.112400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.112589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.112598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.112920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.112930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.113213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.113228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.113536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.113546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.113942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.113953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.114252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.114262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.114545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.114556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.114894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.114905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.115222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.115232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.115440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.115450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.115636] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.115648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.115844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.115854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.116144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.116154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.116444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.116454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.116804] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.116814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.117196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.117206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.117515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.117526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.117869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.117880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.118225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.118234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.118515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.118532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.118843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.118853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.119150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.119160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.048 [2024-11-06 09:08:57.119474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.048 [2024-11-06 09:08:57.119484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.048 qpair failed and we were unable to recover it. 00:33:07.326 [2024-11-06 09:08:57.119769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.326 [2024-11-06 09:08:57.119782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.120098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.120109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.120398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.120407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.120579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.120590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.120913] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.120923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.121288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.121299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.121596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.121606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.121930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.121941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.122244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.122254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.122565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.122575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.122878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.122888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.123197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.123242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.123332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.123342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.123557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.123571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.123794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.123806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.124137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.124147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.124345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.124355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.124675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.124684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.124877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.124887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.125223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.125233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.125542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.125552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.125772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.125782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.126077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.126087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.126462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.126471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.126770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.126780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.127124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.127134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.127441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.127451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.127659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.127670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.127992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.128002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.128309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.128319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.128658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.128667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.128966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.128977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.129213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.129222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.129529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.129540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.129855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.129865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.130155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.130164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.130472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.130482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.130842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.130853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.327 qpair failed and we were unable to recover it. 00:33:07.327 [2024-11-06 09:08:57.131123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.327 [2024-11-06 09:08:57.131132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.131456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.131467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.131638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.131648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.131966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.131976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.132302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.132312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.132614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.132624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.132826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.132837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.133168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.133178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.133370] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.133380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.133696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.133706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.134005] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.134017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.134316] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.134326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.134703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.134713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.135073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.135090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.135393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.135402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.135685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.135701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.136069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.136079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.136451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.136463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.136683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.136693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.136815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.136824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.137089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.137099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.137365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.137374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.137712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.137722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.138104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.138115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.138409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.138420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.138713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.138723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.139029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.139039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.139388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.139397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.139703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.139713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.140024] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.140034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.140323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.140334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.140672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.140683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.140983] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.140994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.141314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.141324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.141603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.141614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.141914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.141924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.142190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.142200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.142526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.142535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.142828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.328 [2024-11-06 09:08:57.142838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.328 qpair failed and we were unable to recover it. 00:33:07.328 [2024-11-06 09:08:57.143146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.143156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.143470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.143480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.143788] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.143798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.144121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.144131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.144301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.144311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.144617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.144632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.145001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.145012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.145296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.145305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.145588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.145599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.145907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.145917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.146195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.146205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.146487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.146496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.146794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.146805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.147127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.147137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.147444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.147454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.147777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.147788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.148065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.148074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.148475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.148485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.148802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.148813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.149116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.149127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.149424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.149433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.149643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.149653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.149927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.149939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.150149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.150159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.150354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.150364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.150668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.150679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.150969] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.150980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.151276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.151285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.151569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.151580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.151898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.151908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.152198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.152208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.152392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.152401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.152733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.152743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.153030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.153040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.153318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.153338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.153652] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.153662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.153976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.153986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.154184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.154193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.329 [2024-11-06 09:08:57.154497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.329 [2024-11-06 09:08:57.154507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.329 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.154816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.154827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.155190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.155200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.155496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.155506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.155817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.155828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.156146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.156156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.156461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.156471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.156785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.156795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.157089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.157102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.157435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.157445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.157737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.157753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.158076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.158086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.158404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.158413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.158693] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.158703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.158987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.158998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.159307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.159317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.159607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.159619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.159952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.159963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.160268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.160278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.160561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.160571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.160913] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.160924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.161275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.161286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.161612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.161623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.161924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.161935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.162215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.162226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.162575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.162585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.162903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.162913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.163226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.163235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.163511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.163521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.163835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.163845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.164163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.164173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.164496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.164506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.164796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.164806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.165015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.165025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.165346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.165355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.165686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.165698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.165886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.165897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.166183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.166193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.166476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.166486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.330 [2024-11-06 09:08:57.166810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.330 [2024-11-06 09:08:57.166821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.330 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.167110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.167119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.167443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.167453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.167775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.167786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.168093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.168103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.168401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.168411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.168612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.168622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.168811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.168822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.169138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.169148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.169462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.169472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.169757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.169767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.170055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.170065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.170408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.170418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.170699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.170708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.171013] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.171023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.171257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.171267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.171486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.171496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.171807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.171818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.172181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.172190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.172488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.172498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.172805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.172815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.173199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.173211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.173521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.173530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.173851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.173861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.174191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.174200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.174593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.174602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.174925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.174936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.175256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.175266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.175653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.175663] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.175997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.176008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.176344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.176353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.176645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.176656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.176844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.176855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.331 qpair failed and we were unable to recover it. 00:33:07.331 [2024-11-06 09:08:57.177191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.331 [2024-11-06 09:08:57.177201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.177485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.177496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.177797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.177808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.178113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.178123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.178338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.178350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.178519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.178530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.178828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.178839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.179124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.179134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.179417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.179428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.179768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.179779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.180171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.180181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.180464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.180474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.180785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.180796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.181093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.181103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.181507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.181518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.181813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.181824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.182037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.182047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.182354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.182365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.182692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.182702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.182909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.182919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.183229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.183239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.183591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.183601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.183798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.183808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.184117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.184127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.184470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.184480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.184806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.184816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.185034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.185043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.185362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.185372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.185692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.185703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.186061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.186071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.186429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.186439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.186741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.186770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.187128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.187138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.187438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.187447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.187769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.187780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.187992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.188002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.188334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.188344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.188672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.188682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.188850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.332 [2024-11-06 09:08:57.188861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.332 qpair failed and we were unable to recover it. 00:33:07.332 [2024-11-06 09:08:57.189145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.189155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.189464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.189474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.189772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.189782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.190061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.190071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.190275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.190285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.190603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.190613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.190909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.190921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.191235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.191244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.191539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.191549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.191717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.191727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.192045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.192055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.192361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.192372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.192677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.192687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.193011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.193021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.193407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.193416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.193737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.193751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.193994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.194004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.194323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.194332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.194662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.194673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.194896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.194908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.195217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.195228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.195535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.195546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.195730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.195741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.196046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.196057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.196363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.196374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.196681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.196692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.197029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.197041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.197341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.197351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.197654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.197665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.198023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.198034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.198359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.198370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.198677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.198687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.198891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.198902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.199117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.199131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.199435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.199445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.199749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.199761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.199835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.199846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.200037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.200047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.333 [2024-11-06 09:08:57.200267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.333 [2024-11-06 09:08:57.200276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.333 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.200533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.200542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.200844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.200854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.201132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.201143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.201469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.201479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.201763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.201773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.202108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.202118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.202423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.202432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.202765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.202775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.203093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.203111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.203414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.203425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.203731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.203741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.204120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.204130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.204436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.204445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.204752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.204762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.205107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.205117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.205421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.205431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.205738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.205752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.206128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.206137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.206329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.206339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.206661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.206670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.206961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.206971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.207268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.207278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.207593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.207604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.207922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.207933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.208245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.208255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.208551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.208560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.208869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.208880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.209194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.209204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.209507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.209516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.209771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.209782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.210119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.210128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.210302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.210313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.210528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.210544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.210853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.210863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.211158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.211169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.211510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.211519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.211898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.211908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.212213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.212223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.334 qpair failed and we were unable to recover it. 00:33:07.334 [2024-11-06 09:08:57.212421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.334 [2024-11-06 09:08:57.212430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.212621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.212630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.212829] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.212839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.212915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.212925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.213232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.213243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.213570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.213580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.213877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.213887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.214204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.214215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.214403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.214414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.214735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.214753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.215062] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.215073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.215377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.215386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.215690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.215700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.215993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.216004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.216305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.216315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.216626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.216637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.216957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.216967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.217349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.217359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.217550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.217560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.217897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.217907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.218223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.218234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.218568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.218578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.218883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.218893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.219199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.219209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.219513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.219525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.219821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.219831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.220138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.220148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.220464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.220473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.220752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.220763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.221180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.221190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.221471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.221480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.221655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.221665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.221878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.221888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.222200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.222210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.222514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.222524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.222918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.222929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.223255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.223265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.223545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.223554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.223848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.223859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.335 qpair failed and we were unable to recover it. 00:33:07.335 [2024-11-06 09:08:57.224159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.335 [2024-11-06 09:08:57.224168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.224472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.224481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.224766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.224777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.225094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.225110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.225415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.225425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.225708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.225719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.226039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.226049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.226354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.226364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.226665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.226676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.226987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.226997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.227376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.227386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.227709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.227720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.228029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.228040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.228366] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.228377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.228713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.228724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.229070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.229082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.229403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.229414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.229720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.229731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.230032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.230042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.230333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.230344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.230523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.230534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.230843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.230853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.231153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.231162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.231469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.231479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.231788] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.231799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.232124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.232133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.232507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.232517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.232821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.232831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.233143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.233152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.233467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.233478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.233798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.233808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.336 [2024-11-06 09:08:57.234123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.336 [2024-11-06 09:08:57.234133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.336 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.234435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.234445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.234754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.234764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.234991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.235001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.235287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.235297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.235609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.235619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.235924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.235935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.236223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.236233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.236580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.236590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.236896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.236906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.237227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.237236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.237445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.237454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.237777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.237787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.238127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.238137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.238424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.238440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.238768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.238779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.239094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.239109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.239426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.239435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.239740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.239755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.240079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.240089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.240389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.240399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.240724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.240734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.240937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.240950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.241268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.241277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.241552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.241562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.241880] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.241890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.242170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.242189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.242483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.242493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.242807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.242817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.243210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.243222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.243456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.243466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.243817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.243827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.244133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.244152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.244470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.244480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.244787] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.244797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.245105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.245114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.245417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.245427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.245738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.245752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.246053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.246063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.337 qpair failed and we were unable to recover it. 00:33:07.337 [2024-11-06 09:08:57.246377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.337 [2024-11-06 09:08:57.246387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.246676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.246688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.247000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.247011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.247310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.247320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.247647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.247657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.247952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.247962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.248263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.248273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.248583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.248593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.248889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.248899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.249202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.249213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.249512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.249522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.249840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.249850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.250150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.250160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.250461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.250471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.250796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.250807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.251117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.251127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.251289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.251300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.251648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.251657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.252052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.252063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.252334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.252344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.252644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.252654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.252973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.252984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.253189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.253198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.253562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.253572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.253850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.253863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.254162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.254171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.254481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.254491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.254807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.254817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.254986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.254997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.255236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.255246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.255547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.255557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.255848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.255859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.256029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.256040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.256332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.256342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.256633] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.256644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.256943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.256953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.257349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.257360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.257703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.257713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.257890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.338 [2024-11-06 09:08:57.257902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.338 qpair failed and we were unable to recover it. 00:33:07.338 [2024-11-06 09:08:57.258187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.258197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.258490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.258509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.258825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.258835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.259171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.259182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.259367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.259377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.259696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.259707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.260025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.260035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.260360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.260369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.260657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.260667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.260860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.260871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.261229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.261239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.261538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.261549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.261881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.261894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.262313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.262323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.262493] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.262504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.262790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.262800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.262968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.262979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.263273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.263282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.263596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.263606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.263912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.263923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.264229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.264240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.264415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.264426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.264803] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.264813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.265080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.265090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.265400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.265410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.265699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.265709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.265931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.265942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.266292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.266302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.266505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.266516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.266825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.266835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.267162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.267172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.267478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.267488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.267693] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.267704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.268055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.268065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.268382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.268391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.268700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.268709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.269127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.269137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.269450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.269460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.269789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.339 [2024-11-06 09:08:57.269799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.339 qpair failed and we were unable to recover it. 00:33:07.339 [2024-11-06 09:08:57.270097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.270114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.270293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.270304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.270618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.270628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.270918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.270929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.271238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.271249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.271555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.271566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.271891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.271901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.272109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.272119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.272343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.272353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.272669] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.272679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.272991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.273001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.273317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.273327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.273607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.273616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.273912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.273922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.274300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.274314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.274642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.274652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.274968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.274979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.275282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.275292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.275505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.275515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.275780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.275790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.275961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.275972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.276230] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.276240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.276521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.276531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.276882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.276893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.277285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.277294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.277609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.277618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.277933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.277943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.278239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.278254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.278568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.278578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.278886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.278896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.279226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.279236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.279483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.279492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.279705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.279716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.279938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.279948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.340 [2024-11-06 09:08:57.280245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.340 [2024-11-06 09:08:57.280255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.340 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.280568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.280578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.280868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.280880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.281185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.281195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.281565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.281576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.281888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.281898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.282216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.282226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.282536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.282548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.282758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.282773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.283071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.283082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.283273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.283283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.283551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.283562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.283761] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.283771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.284074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.284084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.284283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.284292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.284612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.284622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.284918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.284928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.285124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.285135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.285532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.285542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.285835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.285845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.286146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.286156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.286502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.286513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.286836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.286846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.287147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.287158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.287359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.287369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.287521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.287531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.287744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.287769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.288087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.288097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.288449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.288459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.288634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.288643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.289016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.289026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.289295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.289304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.289643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.289653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.289849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.289860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.290202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.290212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.290498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.290508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.290826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.290836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.291183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.291194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.291538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.291549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.341 [2024-11-06 09:08:57.291853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.341 [2024-11-06 09:08:57.291863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.341 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.292190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.292201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.292529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.292539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.292854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.292864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.293175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.293185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.293521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.293531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.293862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.293873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.294176] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.294187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.294489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.294499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.294695] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.294707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.295032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.295042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.295368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.295378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.295570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.295580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.295779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.295789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.296073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.296083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.296398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.296407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.296598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.296609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.296801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.296811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.297123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.297133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.297414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.297423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.297738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.297752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.298090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.298100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.298409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.298418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.298596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.298608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.298920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.298931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.299223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.299232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.299582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.299592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.299880] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.299890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.300211] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.300222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.300514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.300525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.300716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.300726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.301039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.301049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.301357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.301367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.301674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.301685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.301990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.302001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.302204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.302215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.302404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.302417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.302721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.302733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.303079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.303090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.342 [2024-11-06 09:08:57.303369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.342 [2024-11-06 09:08:57.303379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.342 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.303674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.303684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.304007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.304018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.304353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.304364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.304574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.304585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.304772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.304783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.305100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.305109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.305390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.305407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.305714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.305724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.306002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.306012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.306202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.306212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.306536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.306546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.306763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.306773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.307102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.307112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.307439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.307450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.307780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.307791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.308137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.308147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.308420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.308430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.308745] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.308760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.309072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.309082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.309393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.309403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.309719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.309729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.310088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.310098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.310411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.310421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.310725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.310735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.311121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.311132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.311479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.311489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.311797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.311808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.312109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.312120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.312432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.312442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.312631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.312642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.312925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.312935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.313225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.313235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.313544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.313554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.313870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.313881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.314183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.314193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.314507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.314517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.314884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.314895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.315231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.315242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.343 qpair failed and we were unable to recover it. 00:33:07.343 [2024-11-06 09:08:57.315556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.343 [2024-11-06 09:08:57.315574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.315887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.315898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.316193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.316209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.316543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.316553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.316848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.316858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.317191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.317201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.317403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.317413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.317623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.317634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.317814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.317826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.318157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.318167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.318468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.318478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.318791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.318801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.319147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.319157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.319480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.319489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.319806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.319816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.320113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.320124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.320425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.320435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.320735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.320749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.321043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.321053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.321338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.321349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.321645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.321656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.321979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.321989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.322290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.322301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.322631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.322641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.322980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.322991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.323319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.323330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.323604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.323614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.323942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.323952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.324330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.324340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.324617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.324627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.324911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.324921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.325240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.325251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.325528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.325538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.325823] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.325833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.326190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.326200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.326405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.326415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.326753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.326764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.326963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.326972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.327281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.327292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.344 qpair failed and we were unable to recover it. 00:33:07.344 [2024-11-06 09:08:57.327614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.344 [2024-11-06 09:08:57.327624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.327910] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.327921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.328290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.328300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.328587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.328597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.328958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.328968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.329367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.329376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.329697] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.329707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.330002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.330012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.330388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.330398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.330621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.330632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.330934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.330945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.331277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.331287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.331591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.331601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.331939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.331949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.332241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.332259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.332326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.332335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.332579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.332588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.332796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.332806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.333117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.333127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.333535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.333545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.333757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.333767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.333946] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.333956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.334311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.334321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.334520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.334529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.334856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.334867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.335157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.335168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.335373] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.335392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.335699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.335708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.335934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.335946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.336277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.336287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.336675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.336686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.337074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.337084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.337151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.337161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.337454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.337465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.345 qpair failed and we were unable to recover it. 00:33:07.345 [2024-11-06 09:08:57.337704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.345 [2024-11-06 09:08:57.337715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.337917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.337928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.338138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.338149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.338339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.338350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.338658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.338669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.338953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.338964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.339323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.339334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.339384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.339394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.339673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.339684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.339874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.339885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.340254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.340264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.340661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.340671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.341053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.341064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.341404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.341415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.341788] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.341798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.342095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.342105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.342257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.342267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.342602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.342611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.342786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.342798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.343145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.343155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.343425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.343435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.343781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.343792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.344000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.344010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.344325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.344335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.344622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.344632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.344944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.344954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.345282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.345293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.345611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.345621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.345911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.345921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.346247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.346257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.346647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.346657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.346994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.347004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.347327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.347337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.347717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.347728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.348052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.348062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.348372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.348384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.348668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.348677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.348871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.348882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.349244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.346 [2024-11-06 09:08:57.349254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.346 qpair failed and we were unable to recover it. 00:33:07.346 [2024-11-06 09:08:57.349536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.349546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.349864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.349874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.349986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.349996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.350308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.350318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.350611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.350621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.350791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.350802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.350977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.350987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.351157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.351168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.351439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.351449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.351756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.351766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.352165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.352175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.352346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.352357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.352635] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.352645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.352950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.352961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.353277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.353287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.353627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.353637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.353922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.353932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.354275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.354285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.354589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.354600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.354958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.354968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.355266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.355277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.355590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.355600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.355899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.355909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.356209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.356222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.356509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.356519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.356839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.356849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.357190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.357200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.357494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.357504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.357665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.357676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.357763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.357774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.358005] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.358015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.358400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.358410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.358704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.358715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.359013] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.359024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.359389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.359400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.359682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.359692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.359986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.359997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.360304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.360315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.360622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.347 [2024-11-06 09:08:57.360633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.347 qpair failed and we were unable to recover it. 00:33:07.347 [2024-11-06 09:08:57.360963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.360973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.361274] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.361283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.361608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.361617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.361902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.361913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.362200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.362210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.362563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.362573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.362752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.362763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.363050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.363060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.363370] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.363380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.363757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.363768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.364092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.364102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.364410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.364419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.364755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.364766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.365144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.365154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.365483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.365493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.365802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.365813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.366111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.366122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.366443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.366452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.366763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.366773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.366958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.366968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.367271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.367281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.367586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.367596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.367891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.367902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.368216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.368226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.368504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.368514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.368797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.368810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.369120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.369130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.369419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.369429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.369776] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.369786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.370098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.370108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.370424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.370433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.370741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.370755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.371078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.371088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.371396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.371406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.371712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.371723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.372025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.372036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.372368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.372379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.372681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.372692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.348 qpair failed and we were unable to recover it. 00:33:07.348 [2024-11-06 09:08:57.373006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.348 [2024-11-06 09:08:57.373017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.373333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.373344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.373683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.373694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.373994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.374005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.374314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.374325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.374628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.374639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.374827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.374838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.375041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.375051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.375348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.375359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.375664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.375675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.375948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.375959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.376281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.376291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.376489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.376500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.376807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.376817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.377115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.377126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.377323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.377333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.377591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.377601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.377889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.377899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.378190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.378200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.378487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.378497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.378830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.378840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.379139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.379157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.379487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.379496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.379674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.379684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.379993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.380004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.380312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.380322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.380656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.380666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.381008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.381019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.381325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.381335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.381618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.381628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.381791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.381803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.382171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.382182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.382482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.382492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.382802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.382812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.349 [2024-11-06 09:08:57.383096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.349 [2024-11-06 09:08:57.383106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.349 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.383418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.383428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.383651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.383661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.383933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.383943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.384257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.384267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.384625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.384635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.384928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.384939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.385246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.385256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.385576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.385587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.385739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.385759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.386029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.386038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.386345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.386355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.386649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.386659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.386965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.386976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.387284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.387294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.387582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.387593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.387914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.387924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.388224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.388234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.388534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.388544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.388809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.388820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.389140] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.389150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.389464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.389476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.389786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.389796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.390092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.390111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.390403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.390413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.390714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.390725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.391087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.391098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.391406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.391416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.391752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.391763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.392104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.392114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.392396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.392406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.392706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.392716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.393061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.393072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.393384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.393394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.393704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.393714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.394035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.394051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.394393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.394403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.394703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.394713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.395021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.395031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.350 [2024-11-06 09:08:57.395341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.350 [2024-11-06 09:08:57.395352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.350 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.395670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.395680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.395982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.395992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.396302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.396312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.396619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.396629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.396941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.396951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.397268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.397287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.397598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.397608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.397922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.397932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.398223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.398235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.398426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.398437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.398758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.398769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.398939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.398950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.399180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.399190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.399504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.399514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.399799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.399809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.400132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.400142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.400426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.400436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.400749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.400760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.400989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.400998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.401298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.401308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.401593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.401603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.401884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.401895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.402211] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.402221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.402525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.402535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.402768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.402779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.403104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.403114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.403416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.403426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.403721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.403733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.404035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.404046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.404365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.404376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.404693] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.404703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.404998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.405008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.405302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.405311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.405615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.405625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.405916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.405926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.406245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.406255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.406587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.406598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.406904] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.406914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.351 qpair failed and we were unable to recover it. 00:33:07.351 [2024-11-06 09:08:57.407215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.351 [2024-11-06 09:08:57.407225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.407529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.407539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.407922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.407932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.408244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.408253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.408568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.408578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.408886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.408896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.409230] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.409240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.409541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.409551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.409861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.409872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.410205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.410216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.410417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.410427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.410703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.410715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.411027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.411038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.411339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.411349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.411649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.411659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.411944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.411954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.412273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.412283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.412583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.412593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.412879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.412889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.413206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.413216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.413517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.413527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.413893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.413904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.414135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.414144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.414459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.414469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.414775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.414786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.415113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.415123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.415497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.415506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.415813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.415823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.416127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.416137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.416442] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.416452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.416782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.416792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.417067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.417077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.417385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.417394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.417706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.417716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.418028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.418038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.418310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.418320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.418538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.418548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.418894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.418904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.419233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.419243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.352 qpair failed and we were unable to recover it. 00:33:07.352 [2024-11-06 09:08:57.419550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.352 [2024-11-06 09:08:57.419560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.353 qpair failed and we were unable to recover it. 00:33:07.353 [2024-11-06 09:08:57.419860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.353 [2024-11-06 09:08:57.419870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.353 qpair failed and we were unable to recover it. 00:33:07.353 [2024-11-06 09:08:57.420184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.353 [2024-11-06 09:08:57.420194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.353 qpair failed and we were unable to recover it. 00:33:07.353 [2024-11-06 09:08:57.420476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.353 [2024-11-06 09:08:57.420485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.353 qpair failed and we were unable to recover it. 00:33:07.353 [2024-11-06 09:08:57.420674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.353 [2024-11-06 09:08:57.420683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.353 qpair failed and we were unable to recover it. 00:33:07.353 [2024-11-06 09:08:57.421010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.353 [2024-11-06 09:08:57.421020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.353 qpair failed and we were unable to recover it. 00:33:07.353 [2024-11-06 09:08:57.421321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.353 [2024-11-06 09:08:57.421330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.353 qpair failed and we were unable to recover it. 00:33:07.353 [2024-11-06 09:08:57.421609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.353 [2024-11-06 09:08:57.421619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.353 qpair failed and we were unable to recover it. 00:33:07.353 [2024-11-06 09:08:57.421914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.353 [2024-11-06 09:08:57.421924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.353 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.422238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.422250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.422555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.422565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.422855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.422865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.423193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.423203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.423505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.423516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.423824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.423835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.424135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.424145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.424446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.424456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.424740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.424756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.424972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.424984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.425283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.425293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.425603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.425613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.425917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.425928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.426241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.426251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.426538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.426547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.426872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.426883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.427186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.427196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.427498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.427507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.427794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.427804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.428122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.428133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.428441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.428452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.428758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.428769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.429079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.429090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.429393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.429403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.429709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.429720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.429869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.429881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.430184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.430195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.430488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.430498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.430800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.430810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.431122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.431132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.431414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.431424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.431768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.431781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.432044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.432053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.432218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.432229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.631 [2024-11-06 09:08:57.432533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.631 [2024-11-06 09:08:57.432543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.631 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.432866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.432877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.433191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.433201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.433482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.433499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.433830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.433840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.434145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.434155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.434336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.434346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.434585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.434595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.434885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.434896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.435198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.435209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.435510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.435521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.435826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.435837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.436139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.436149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.436446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.436456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.436757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.436767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.436965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.436975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.437286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.437297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.437555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.437565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.437878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.437888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.438107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.438117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.438458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.438468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.438754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.438764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.439051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.439061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.439340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.439350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.439571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.439581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.439896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.439908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.440221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.440231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.440538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.440548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.440847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.440858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.441200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.441211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.441493] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.441503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.441852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.441862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.442141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.442151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.442460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.442470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.442756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.442766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.442939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.442949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.443255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.443265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.443583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.443593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.443827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.443837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.632 qpair failed and we were unable to recover it. 00:33:07.632 [2024-11-06 09:08:57.444159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.632 [2024-11-06 09:08:57.444168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.444470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.444481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.444683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.444694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.445001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.445011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.445242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.445252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.445559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.445569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.445877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.445887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.446180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.446190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.446492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.446503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.446832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.446842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.447177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.447187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.447533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.447543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.447757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.447768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.447972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.447981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.448315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.448325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.448677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.448687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.448893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.448904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.449215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.449226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.449560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.449570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.449878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.449888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.450177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.450187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.450519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.450528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.450736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.450750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.451066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.451076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.451381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.451391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.451718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.451729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.452050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.452063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.452379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.452388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.452696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.452706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.453113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.453123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.453411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.453421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.453614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.453624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.453900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.453910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.454104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.454114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.454419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.454428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.454613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.454622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.454884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.454894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.455166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.455175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.455366] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.455377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.633 qpair failed and we were unable to recover it. 00:33:07.633 [2024-11-06 09:08:57.455706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.633 [2024-11-06 09:08:57.455717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.456114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.456125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.456442] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.456451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.456760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.456771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.457142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.457153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.457467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.457476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.457611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.457621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.457961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.457971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.458147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.458158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.458459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.458470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.458787] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.458798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.459110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.459120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.459422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.459432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.459721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.459731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.460039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.460050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.460335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.460345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.460614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.460624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.460969] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.460979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.461156] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.461167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.461491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.461501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.461686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.461696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.462059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.462070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.462394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.462404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.462690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.462700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.463020] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.463030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.463344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.463362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.463692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.463701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.464003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.464013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.464323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.464336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.464661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.464680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.464999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.465010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.465305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.465315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.465631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.465641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.465953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.465963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.466246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.466256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.466563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.466573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.466879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.466890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.467173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.467189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.467521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.467532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.634 qpair failed and we were unable to recover it. 00:33:07.634 [2024-11-06 09:08:57.467846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.634 [2024-11-06 09:08:57.467858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.468083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.468093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.468409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.468419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.468641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.468651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.469058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.469069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.469401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.469410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.469717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.469726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.470013] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.470023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.470343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.470353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.470630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.470640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.471077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.471087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.471377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.471388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.471694] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.471704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.472003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.472016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.472293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.472303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.472582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.472592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.472905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.472917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.473205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.473216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.473545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.473556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.473866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.473877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.474189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.474200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.474506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.474516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.474894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.474904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.475212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.475222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.475417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.475427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.475793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.475803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.476095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.476105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.476381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.476391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.476697] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.476707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.477015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.477025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.477344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.477354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.477693] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.477704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.477988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.635 [2024-11-06 09:08:57.477999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.635 qpair failed and we were unable to recover it. 00:33:07.635 [2024-11-06 09:08:57.478324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.478335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.478638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.478649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.478941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.478951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.479266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.479275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.479584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.479594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.479914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.479923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.480129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.480138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.480446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.480456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.480763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.480773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.481081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.481091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.481380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.481390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.481672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.481682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.481973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.481983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.482302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.482312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.482631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.482642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.482911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.482922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.483211] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.483221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.483524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.483534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.483897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.483907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.484223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.484232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.484504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.484514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.484791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.484801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.485121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.485131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.485447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.485457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.485761] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.485774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.486075] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.486084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.486371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.486381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.486570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.486580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.486847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.486857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.487195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.487206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.487495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.487505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.487711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.487721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.488039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.488049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.488363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.488373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.488701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.488710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.488993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.489003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.489306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.489315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.489509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.489519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.489851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.636 [2024-11-06 09:08:57.489862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.636 qpair failed and we were unable to recover it. 00:33:07.636 [2024-11-06 09:08:57.490057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.490067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.490299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.490310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.490618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.490628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.490920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.490932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.491288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.491299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.491578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.491588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.491901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.491911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.492184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.492194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.492520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.492530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.492806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.492816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.493145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.493155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.493416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.493427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.493757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.493771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.494081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.494091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.494421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.494431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.494755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.494766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.495174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.495184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.495446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.495455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.495763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.495774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.496151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.496162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.496335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.496346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.496674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.496684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.496977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.496987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.497301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.497319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.497516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.497526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.497809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.497820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.498144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.498154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.498337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.498347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.498734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.498744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.499045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.499056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.499388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.499398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.499703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.499714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.500031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.500042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.500350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.500359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.500684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.500693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.501002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.501017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.501331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.501341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.501722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.501732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.502040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.637 [2024-11-06 09:08:57.502050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.637 qpair failed and we were unable to recover it. 00:33:07.637 [2024-11-06 09:08:57.502217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.502228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.502597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.502607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.502885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.502896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.503215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.503225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.503536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.503546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.503835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.503845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.504153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.504163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.504472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.504483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.504839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.504849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.505050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.505060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.505408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.505417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.505837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.505848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.506141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.506151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.506446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.506457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.506786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.506799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.506965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.506976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.507333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.507343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.507668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.507678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.507986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.507996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.508294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.508303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.508620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.508630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.508920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.508930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.509234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.509244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.509535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.509545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.509867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.509877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.510215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.510225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.510510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.510519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.510692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.510703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.510995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.511006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.511332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.511343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.511664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.511675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.511990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.512001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.512375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.512385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.512634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.512644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.512975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.512985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.513173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.513183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.513483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.513493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.513837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.513848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.514028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.638 [2024-11-06 09:08:57.514038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.638 qpair failed and we were unable to recover it. 00:33:07.638 [2024-11-06 09:08:57.514411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.514421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.514834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.514845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.515139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.515150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.515324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.515333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.515643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.515653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.515971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.515982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.516412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.516422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.516703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.516713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.516993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.517003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.517118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.517127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.517390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.517401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.517700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.517711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.518084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.518095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.518429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.518439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.518711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.518720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.519023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.519034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.519347] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.519357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.519570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.519579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.519878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.519889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.520210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.520221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.520539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.520557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.520866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.520876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.521050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.521060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.521271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.521281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.521598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.521607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.522000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.522009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.522381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.522392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.522721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.522731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.523080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.523090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.523317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.523326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.523657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.523668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.523964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.523975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.524279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.524290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.524600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.524611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.524923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.524934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.525236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.525248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.525568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.525578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.525873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.525883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.526208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.639 [2024-11-06 09:08:57.526218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.639 qpair failed and we were unable to recover it. 00:33:07.639 [2024-11-06 09:08:57.526529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.526540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.526870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.526881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.527259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.527270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.527579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.527589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.527905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.527917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.528247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.528257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.528456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.528466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.528648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.528657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.528870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.528880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.529097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.529107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.529353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.529363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.529650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.529659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.530060] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.530071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.530343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.530352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.530725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.530735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.530938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.530949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.531254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.531264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.531548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.531566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.531910] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.531921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.532141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.532151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.532317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.532327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.532532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.532543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.532843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.532853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.533162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.533173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.533484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.533494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.533792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.533802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.534111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.534121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.534450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.534460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.534791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.534801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.534995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.535006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.535302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.535312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.535649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.535659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.535955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.535966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.536178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.536188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.640 qpair failed and we were unable to recover it. 00:33:07.640 [2024-11-06 09:08:57.536385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.640 [2024-11-06 09:08:57.536396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.536755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.536765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.537067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.537077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.537401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.537410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.537703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.537714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.538030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.538040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.538237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.538248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.538428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.538439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.538762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.538774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.539099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.539109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.539424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.539434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.539740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.539755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.540048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.540059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.540429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.540439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.540800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.540811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.541108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.541118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.541419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.541428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.541736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.541750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.542043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.542054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.542341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.542351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.542744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.542759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.543066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.543075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.543260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.543270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.543475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.543484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.543790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.543801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.544177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.544187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.544459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.544469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.544635] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.544646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.544951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.544962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.545144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.545155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.545445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.545455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.545782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.545792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.546126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.546136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.546434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.546450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.546704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.546715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.547028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.547039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.547264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.547275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.547483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.547494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.547688] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.547701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.641 qpair failed and we were unable to recover it. 00:33:07.641 [2024-11-06 09:08:57.547889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.641 [2024-11-06 09:08:57.547900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.548094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.548105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.548401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.548412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.548737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.548752] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.548956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.548966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.549344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.549355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.549662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.549672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.549955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.549965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.550278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.550288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.550667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.550677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.550849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.550858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.551164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.551175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.551483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.551493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.551667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.551677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.551955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.551972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.552271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.552281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.552593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.552602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.552800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.552810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.553152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.553161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.553500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.553511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.553814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.553824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.554158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.554168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.554348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.554357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.554691] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.554702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.555031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.555042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.555348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.555357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.555675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.555685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.556042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.556053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.556384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.556394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.556750] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.556760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.557088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.557098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.557425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.557436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.557604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.557615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.557843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.557853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.558188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.558199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.558410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.558419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.558603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.558613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.558792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.558802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.559135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.642 [2024-11-06 09:08:57.559145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.642 qpair failed and we were unable to recover it. 00:33:07.642 [2024-11-06 09:08:57.559468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.559478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.559792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.559804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.560114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.560124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.560495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.560506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.560863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.560874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.561194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.561203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.561515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.561525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.561837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.561847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.562167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.562183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.562498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.562508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.562693] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.562704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.563030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.563041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.563341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.563352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.563519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.563529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.563813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.563823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.564190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.564200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.564491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.564507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.564830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.564840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.565235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.565245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.565527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.565537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.565873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.565884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.566213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.566223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.566476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.566487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.566785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.566796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.566968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.566978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.567337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.567347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.567651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.567661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.567851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.567861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.568076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.568088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.568456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.568466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.568749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.568759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.568945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.568955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.569277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.569287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.569471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.569481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.569839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.569850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.570185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.570195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.570525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.570535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.570824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.570834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.643 [2024-11-06 09:08:57.571168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.643 [2024-11-06 09:08:57.571179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.643 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.571460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.571470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.571668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.571678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.572090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.572101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.572425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.572435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.572737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.572751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.573069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.573079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.573462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.573472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.573534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.573544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.573893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.573904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.574213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.574223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.574540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.574550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.574871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.574881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.575179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.575189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.575512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.575523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.575721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.575731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.576067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.576078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.576371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.576381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.576702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.576712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.577015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.577027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.577343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.577354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.577687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.577697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.577983] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.577994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.578328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.578340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.578647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.578658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.578973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.578984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.579286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.579296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.579600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.579611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.579899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.579910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.580236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.580246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.580534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.580545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.580863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.580878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.581181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.581191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.581458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.581468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.644 [2024-11-06 09:08:57.581797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.644 [2024-11-06 09:08:57.581808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.644 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.582132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.582142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.582455] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.582465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.582752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.582763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.583148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.583158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.583451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.583461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.583685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.583695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.584017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.584028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.584219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.584229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.584548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.584559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.584878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.584888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.585232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.585241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.585545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.585555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.585856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.585867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.586187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.586196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.586475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.586484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.586681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.586691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.587028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.587039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.587356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.587366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.587647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.587657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.587923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.587933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.588223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.588233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.588540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.588550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.588843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.588853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.589027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.589040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.589361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.589372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.589681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.589691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.589989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.589999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.590303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.590312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.590629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.590640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.590948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.590958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.591290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.591301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.591611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.591622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.591961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.591972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.592280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.592290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.592625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.592635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.592844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.592854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.593175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.593185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.593500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.593519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.645 [2024-11-06 09:08:57.593847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.645 [2024-11-06 09:08:57.593858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.645 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.594163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.594173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.594479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.594489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.594792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.594802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.595097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.595106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.595293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.595303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.595622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.595632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.595961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.595971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.596257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.596267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.596574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.596583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.596875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.596885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.597058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.597069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.597271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.597281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.597624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.597634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.597928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.597939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.598137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.598148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.598452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.598462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.598757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.598767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.599061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.599071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.599376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.599385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.599579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.599589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.599859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.599869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.600050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.600061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.600323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.600332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.600668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.600678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.600987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.600997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.601381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.601394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.601697] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.601707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.602002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.602012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.602200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.602210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.602566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.602577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.602879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.602889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.603189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.603199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.603504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.603515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.603708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.603717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.603940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.603950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.604360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.604371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.604587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.604597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.604915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.604925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.646 [2024-11-06 09:08:57.605238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.646 [2024-11-06 09:08:57.605248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.646 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.605496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.605506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.605892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.605902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.606229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.606239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.606586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.606596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.606900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.606911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.607117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.607127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.607306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.607316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.607623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.607634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.607970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.607981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.608317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.608327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.608636] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.608646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.608981] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.608992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.609286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.609304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.609615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.609627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.609833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.609844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.610122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.610132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.610422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.610432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.610739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.610754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.611087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.611097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.611287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.611297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.611640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.611651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.611963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.611973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.612279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.612289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.612573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.612588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.612928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.612938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.613239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.613250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.613564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.613574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.613878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.613888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.614186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.614196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.614480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.614498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.614805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.614816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.615127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.615137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.615470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.615479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.615785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.615795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.616117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.616127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.616411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.616420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.616612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.616623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.616940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.616950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.647 qpair failed and we were unable to recover it. 00:33:07.647 [2024-11-06 09:08:57.617234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.647 [2024-11-06 09:08:57.617245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.617544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.617554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.617856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.617868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.618218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.618228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.618514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.618523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.618872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.618882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.619085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.619095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.619382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.619392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.619717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.619728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.620080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.620090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.620460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.620470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.620780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.620791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.621092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.621102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.621405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.621415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.621589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.621600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.621872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.621882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.622096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.622109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.622415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.622425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.622720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.622738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.623068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.623078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.623261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.623271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.623488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.623498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.623795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.623805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.624135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.624145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.624337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.624347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.624645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.624655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.625004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.625014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.625318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.625334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.625701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.625711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.626081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.626091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.626426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.626436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.626753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.626764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.626972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.626983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.627158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.627168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.627479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.627489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.627791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.627802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.627986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.627997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.628246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.628256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.628624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.628635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.648 qpair failed and we were unable to recover it. 00:33:07.648 [2024-11-06 09:08:57.628817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.648 [2024-11-06 09:08:57.628828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.629004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.629014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.629210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.629220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.629519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.629529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.629849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.629860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.630172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.630182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.630527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.630537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.630729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.630739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.631031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.631042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.631373] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.631383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.631671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.631682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.631855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.631866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.632184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.632194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.632506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.632516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.632904] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.632914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.633259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.633268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.633593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.633603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.633914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.633925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.634239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.634249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.634454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.634464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.634757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.634768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.635068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.635079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.635467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.635476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.635782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.635793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.636112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.636121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.636313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.636324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.636670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.636680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.636976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.636986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.637185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.637195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.637500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.637511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.637813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.637823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.638139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.638148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.638432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.638447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.638816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.638826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.639135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.639146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.649 qpair failed and we were unable to recover it. 00:33:07.649 [2024-11-06 09:08:57.639441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.649 [2024-11-06 09:08:57.639451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.639775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.639785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.639995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.640005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.640313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.640323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.640649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.640658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.640979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.640989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.641284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.641293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.641601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.641612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.641892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.641902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.642218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.642228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.642530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.642545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.642862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.642872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.643202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.643212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.643514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.643523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.643821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.643831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.644148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.644159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.644440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.644450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.644762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.644772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.645093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.645103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.645413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.645422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.645741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.645756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.646130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.646140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.646467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.646477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.646792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.646803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.647091] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.647100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.647405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.647414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.647728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.647750] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.648078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.648088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.648463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.648473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.648781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.648791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.649116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.649126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.649429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.649439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.649727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.649737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.650044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.650054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.650337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.650347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.650529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.650540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.650896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.650907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.651226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.651236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.651550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.651560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.650 qpair failed and we were unable to recover it. 00:33:07.650 [2024-11-06 09:08:57.651868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.650 [2024-11-06 09:08:57.651878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.652210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.652219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.652419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.652428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.652720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.652730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.653070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.653081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.653374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.653384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.653567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.653577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.653893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.653903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.654215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.654226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.654558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.654567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.654877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.654887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.655207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.655217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.655532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.655545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.655880] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.655890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.656199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.656209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.656509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.656519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.656828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.656838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.657154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.657164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.657462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.657480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.657796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.657807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.658121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.658131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.658410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.658420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.658632] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.658643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.658990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.659001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.659313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.659324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.659651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.659661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.659980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.659990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.660197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.660207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.660526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.660536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.660822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.660832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.661138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.661148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.661452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.661462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.661769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.661780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.662117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.662127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.662433] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.662442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.662792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.662803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.663117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.663127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.663302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.663311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.663634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.663643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.651 qpair failed and we were unable to recover it. 00:33:07.651 [2024-11-06 09:08:57.663951] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.651 [2024-11-06 09:08:57.663965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.664278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.664295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.664623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.664632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.664836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.664846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.665160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.665170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.665497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.665507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.665701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.665712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.666031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.666041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.666260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.666270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.666604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.666614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.666932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.666943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.667261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.667270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.667568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.667578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.667779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.667789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.668110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.668120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.668427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.668436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.668749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.668759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.669140] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.669150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.669491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.669501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.669805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.669815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.670161] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.670171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.670483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.670493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.670872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.670883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.671193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.671203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.671516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.671526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.671842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.671853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.672165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.672174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.672481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.672492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.672801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.672812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.673022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.673031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.673334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.673344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.673664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.673674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.673856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.673866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.674081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.674091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.674403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.674413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.674721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.674731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.675028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.675038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.675324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.675340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.675710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.675719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.652 qpair failed and we were unable to recover it. 00:33:07.652 [2024-11-06 09:08:57.676002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.652 [2024-11-06 09:08:57.676013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.676349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.676359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.676663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.676675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.676883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.676894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.677076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.677085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.677409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.677419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.677735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.677749] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.678088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.678097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.678281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.678291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.678647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.678658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.678943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.678954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.679248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.679258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.679568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.679579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.679890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.679900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.680205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.680215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.680541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.680550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.680850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.680861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.681188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.681199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.681485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.681494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.681779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.681790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.682105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.682114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.682407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.682416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.682717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.682727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.683104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.683114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.683424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.683434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.683763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.683774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.684088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.684099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.684379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.684389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.684700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.684709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.685025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.685038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.685354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.685366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.685667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.685678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.685986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.685998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.686302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.686312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.653 [2024-11-06 09:08:57.686472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.653 [2024-11-06 09:08:57.686483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.653 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.686817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.686828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.687139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.687149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.687434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.687450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.687763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.687774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.687941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.687950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.688310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.688320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.688540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.688550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.688727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.688737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.689058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.689068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.689371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.689380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.689555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.689565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.689861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.689871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.690204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.690214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.690508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.690518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.690827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.690837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.691142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.691151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.691478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.691488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.691794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.691804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.692143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.692154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.692439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.692449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.692729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.692739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.693106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.693115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.693448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.693458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.693760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.693771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.693941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.693950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.694233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.694243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.694567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.694577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.694881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.694891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.695215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.695225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.695508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.695525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.695840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.695850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.696154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.696163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.696458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.696468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.696781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.696790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.697092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.697102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.697284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.697319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.697660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.697670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.697977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.697988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.654 qpair failed and we were unable to recover it. 00:33:07.654 [2024-11-06 09:08:57.698208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.654 [2024-11-06 09:08:57.698218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.698523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.698533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.698820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.698837] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.699049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.699060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.699236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.699247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.699565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.699575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.699919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.699929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.700257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.700268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.700598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.700608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.700919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.700929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.701135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.701146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.701470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.701480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.701866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.701877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.702184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.702194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.702483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.702492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.702801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.702811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.703129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.703140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.703372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.703382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.703692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.703703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.704014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.704026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.704308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.704323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.704648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.704659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.704969] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.704980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.705284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.705293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.705588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.705602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.705786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.705796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.706078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.706087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.706287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.706297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.706613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.706624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.706844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.706854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.707141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.707150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.707479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.707489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.707820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.707831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.708142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.708153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.708487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.708498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.708804] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.708814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.709107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.709117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.709285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.709294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.709574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.709584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.655 qpair failed and we were unable to recover it. 00:33:07.655 [2024-11-06 09:08:57.709876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.655 [2024-11-06 09:08:57.709886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.710208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.710219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.710601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.710611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.710922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.710932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.711147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.711157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.711455] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.711471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.711779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.711789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.712085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.712100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.712413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.712423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.712738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.712754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.713068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.713078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.713369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.713379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.713595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.713606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.713847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.713858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.714255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.714265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.714547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.714557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.714780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.714791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.715104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.715115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.715422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.715432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.715723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.715733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.716073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.716083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.716392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.716402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.716710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.716720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.717086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.717097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.717459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.717470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.717676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.717687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.718004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.718017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.718348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.718357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.718641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.718657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.718991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.719002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.719286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.719303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.719631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.719641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.719960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.719970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.720280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.720290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.720640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.720650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.720867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.720878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.721201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.721211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.721526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.721537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.721859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.656 [2024-11-06 09:08:57.721870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.656 qpair failed and we were unable to recover it. 00:33:07.656 [2024-11-06 09:08:57.722195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.722204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.722394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.722405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.722717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.722727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.723116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.723126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.723416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.723426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.723607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.723616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.723962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.723972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.724287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.724297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.724592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.724602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.724909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.724920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.725213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.725222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.725420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.725431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.725607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.725619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.725924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.725934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.657 [2024-11-06 09:08:57.726221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.657 [2024-11-06 09:08:57.726231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.657 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.726549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.726562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.726848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.726859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.727168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.727178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.727461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.727478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.727791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.727802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.728091] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.728102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.728409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.728419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.728712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.728722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.728905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.728916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.729228] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.729238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.729504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.729514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.729798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.729808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.730107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.730117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.730467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.730477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.730654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.730664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.731102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.731116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.938 qpair failed and we were unable to recover it. 00:33:07.938 [2024-11-06 09:08:57.731281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.938 [2024-11-06 09:08:57.731294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.731654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.731664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.731975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.731986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.732293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.732303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.732635] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.732645] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.732843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.732854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.733078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.733088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.733420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.733430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.733595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.733605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.734026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.734037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.734210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.734221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.734435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.734446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.734768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.734779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.735048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.735059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.735365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.735374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.735682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.735692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.735914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.735925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.736238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.736248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.736556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.736565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.736911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.736922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.737232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.737242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.737449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.737458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.737786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.737796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.738077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.738087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.738300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.738312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.738601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.738611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.939 [2024-11-06 09:08:57.738909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.939 [2024-11-06 09:08:57.738919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.939 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.739239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.739250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.739564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.739574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.739844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.739855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.740161] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.740171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.740396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.740406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.740603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.740613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.740953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.740963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.741276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.741286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.741589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.741600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.741909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.741919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.742238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.742248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.742488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.742499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.742811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.742821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.743134] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.743144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.743425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.743435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.743743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.743757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.744048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.744058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.744442] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.744451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.744688] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.744698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.745002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.745012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.745197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.745207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.745525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.745535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.745821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.745831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.746127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.746144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.746459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.746469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.746754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.746764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.746980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.940 [2024-11-06 09:08:57.746989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.940 qpair failed and we were unable to recover it. 00:33:07.940 [2024-11-06 09:08:57.747300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.747310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.747621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.747631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.747960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.747970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.748187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.748197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.748395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.748405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.748727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.748737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.749039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.749049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.749332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.749341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.749625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.749636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.749955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.749966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.750271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.750280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.750585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.750595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.750785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.750795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.751131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.751141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.751369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.751378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.751706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.751716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.752029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.752039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.752343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.752354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.752657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.752666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.752971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.752981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.753275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.753286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.753474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.753484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.753790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.753801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.754086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.754096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.754414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.754423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.754753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.754764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.941 [2024-11-06 09:08:57.755086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.941 [2024-11-06 09:08:57.755096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.941 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.755382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.755392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.755611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.755622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.755895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.755906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.756233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.756244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.756587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.756597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.756772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.756782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.756945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.756956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.757327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.757337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.757638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.757649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.757990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.758002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.758346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.758357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.758667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.758680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.758992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.759004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.759329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.759340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.759653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.759664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.760003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.760014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.760343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.760353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.760662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.760673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.760992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.761003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.761307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.761318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.761649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.761659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.762023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.762034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.762372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.762383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.762591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.762602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.762956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.762967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.763275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.763285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.942 [2024-11-06 09:08:57.763590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.942 [2024-11-06 09:08:57.763601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.942 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.763915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.763925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.764217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.764227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.764445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.764455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.764774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.764785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.765066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.765076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.765399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.765408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.765619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.765628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.765961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.765972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.766261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.766271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.766560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.766570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.766887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.766898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.767223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.767233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.767504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.767514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.767853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.767863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.768121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.768131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.768470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.768480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.768763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.768773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.769111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.769122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.769343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.769353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.769663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.769674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.769995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.770006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.770294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.770304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.770625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.770635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.770914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.770924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.771304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.771313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.771593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.771605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.943 [2024-11-06 09:08:57.771816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.943 [2024-11-06 09:08:57.771827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.943 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.772226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.772236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.772536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.772547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.772724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.772733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.773045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.773055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.773413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.773423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.773728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.773739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.774050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.774060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.774345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.774362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.774666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.774676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.774872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.774882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.775108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.775118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.775471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.775481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.775819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.775830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.776130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.776140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.776461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.776471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.776764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.776775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.777186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.777197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.777503] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.777512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.777697] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.777707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.778025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.778035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.778360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.778370] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.944 qpair failed and we were unable to recover it. 00:33:07.944 [2024-11-06 09:08:57.778683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.944 [2024-11-06 09:08:57.778693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.778813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.778823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.779198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.779208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.779382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.779391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.779777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.779790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.780098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.780108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.780412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.780422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.780602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.780612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.780923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.780933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.781108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.781118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.781334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.781343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.781623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.781633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.781964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.781981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.782297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.782308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.782634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.782644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.782835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.782846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.783034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.783047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.783435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.783446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.783753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.783764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.783950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.783960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.784299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.784309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.784507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.784516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.784720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.784730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.784938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.784949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.785131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.785141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.785361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.785371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.785704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.785714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.945 [2024-11-06 09:08:57.786000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.945 [2024-11-06 09:08:57.786011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.945 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.786400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.786411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.786582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.786592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.786691] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.786700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.786898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.786909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.787196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.787206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.787540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.787550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.787888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.787898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.788122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.788132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.788428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.788438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.788767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.788778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.789096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.789105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.789305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.789315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.789625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.789636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.789964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.789974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.790276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.790286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.790577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.790586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.790975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.790985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.791310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.791323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.791631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.791641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.791845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.791855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.792069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.792080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.792456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.792467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.792780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.792791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.793009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.793019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.793321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.793331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.793668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.793678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.946 [2024-11-06 09:08:57.793977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.946 [2024-11-06 09:08:57.793989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.946 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.794279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.794289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.794605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.794615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.794917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.794928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.795245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.795255] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.795558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.795569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.795884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.795895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.796266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.796276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.796466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.796476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.796702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.796712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.796914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.796924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.797219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.797229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.797632] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.797642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.798001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.798011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.798203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.798213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.798495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.798506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.798800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.798810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.799103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.799114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.799438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.799450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.799771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.799782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.800099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.800109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.800299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.800310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.800699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.800709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.800895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.800906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.801231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.801241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.801566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.801576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.801881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.801891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.947 qpair failed and we were unable to recover it. 00:33:07.947 [2024-11-06 09:08:57.802201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.947 [2024-11-06 09:08:57.802211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.802528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.802538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.802817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.802828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.803028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.803038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.803362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.803373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.803677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.803687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.803990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.804001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.804306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.804316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.804662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.804673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.805006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.805017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.805165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.805174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.805476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.805486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.805831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.805842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.806047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.806057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.806342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.806352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.806623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.806633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.806914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.806925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.807244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.807254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.807452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.807462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.807774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.807784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.808147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.808157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.808329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.808339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.808644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.808654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.808842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.808852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.809162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.809172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.809459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.809469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.809789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.948 [2024-11-06 09:08:57.809799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.948 qpair failed and we were unable to recover it. 00:33:07.948 [2024-11-06 09:08:57.810164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.810174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.810483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.810493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.810807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.810818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.811138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.811148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.811486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.811496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.811791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.811804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.812117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.812127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.812434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.812444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.812782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.812792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.813139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.813150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.813461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.813471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.813838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.813849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.814163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.814172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.814481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.814492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.814812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.814822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.815173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.815183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.815369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.815380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.815692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.815703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.816040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.816050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.816383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.816394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.816702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.816712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.817018] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.817029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.817355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.817365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.817739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.817754] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.818070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.818079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.818401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.818411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.818707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.949 [2024-11-06 09:08:57.818717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.949 qpair failed and we were unable to recover it. 00:33:07.949 [2024-11-06 09:08:57.819020] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.819031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.819346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.819356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.819687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.819697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.820014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.820025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.820395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.820405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.820711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.820724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.821031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.821042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.821361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.821371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.821640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.821651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.821937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.821947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.822140] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.822151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.822365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.822376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.822565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.822575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.822761] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.822772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.823092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.823102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.823302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.823311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.823647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.823657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.823966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.823976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.824263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.824274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.824438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.824448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.824732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.824742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.825047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.825057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.825378] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.825389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.825651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.825660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.826025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.826035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.826343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.826353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.826628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.826638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.950 qpair failed and we were unable to recover it. 00:33:07.950 [2024-11-06 09:08:57.826814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.950 [2024-11-06 09:08:57.826825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.827166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.827175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.827352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.827361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.827475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.827485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.827798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.827809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.827911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.827920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.828024] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.828034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.828234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.828244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.828544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.828554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.828856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.828867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.829203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.829213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.829528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.829539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.829777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.829787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.830072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.830082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.830393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.830403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.830744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.830758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.831093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.831103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.831427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.831437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.831626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.831636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.831962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.831975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.832190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.832200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.832494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.832504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.951 [2024-11-06 09:08:57.832702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.951 [2024-11-06 09:08:57.832711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.951 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.833022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.833032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.833141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.833151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.833418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.833427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.833586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.833596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.833880] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.833890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.834195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.834205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.834525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.834535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.834826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.834836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.835144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.835154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.835469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.835479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.835663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.835674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.835897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.835907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.836071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.836081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.836427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.836437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.836753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.836764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.836954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.836965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.837306] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.837316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.837622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.837632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.837917] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.837927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.838235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.838246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.838550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.838560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.838732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.838743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.838952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.838961] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.839271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.839280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.839596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.839606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.839899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.839910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.952 qpair failed and we were unable to recover it. 00:33:07.952 [2024-11-06 09:08:57.840086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.952 [2024-11-06 09:08:57.840096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.840416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.840426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.840741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.840755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.841059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.841069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.841358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.841369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.841543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.841553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.841847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.841857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.842234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.842245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.842551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.842561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.842882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.842892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.843301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.843312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.843619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.843630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.844013] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.844024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.844368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.844378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.844583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.844593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.844958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.844968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.845307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.845317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.845672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.845682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.845856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.845866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.846176] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.846186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.846507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.846517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.846833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.846843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.847175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.847185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.847479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.847489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.847671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.847680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.953 [2024-11-06 09:08:57.848011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.953 [2024-11-06 09:08:57.848022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.953 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.848204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.848214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.848396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.848406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.848731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.848741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.849029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.849040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.849345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.849355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.849651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.849667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.849991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.850002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.850377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.850388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.850721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.850732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.851057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.851067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.851391] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.851401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.851730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.851740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.852079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.852092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.852428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.852438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.852754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.852765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.853051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.853061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.853345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.853355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.853690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.853700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.854021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.854040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.854371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.854380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.854670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.854680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.855010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.855020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.855322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.855331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.855670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.855680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.856019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.856030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.856319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.954 [2024-11-06 09:08:57.856330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.954 qpair failed and we were unable to recover it. 00:33:07.954 [2024-11-06 09:08:57.856608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.856618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.856931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.856941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.857231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.857241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.857297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.857308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.857650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.857660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.857825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.857836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.858081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.858091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.858418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.858428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.858612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.858622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.858932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.858942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.859122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.859131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.859416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.859427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.859729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.859739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.859913] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.859924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.860244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.860253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.860565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.860575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.860778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.860788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.861107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.861117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.861459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.861469] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.861809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.861820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.862003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.862013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.862339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.862348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.862643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.862652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.863021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.863031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.863379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.863388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.863697] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.863707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.955 qpair failed and we were unable to recover it. 00:33:07.955 [2024-11-06 09:08:57.864086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.955 [2024-11-06 09:08:57.864097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.864364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.864376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.864687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.864696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.865074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.865085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.865284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.865294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.865559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.865569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.865875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.865886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.866205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.866215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.866496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.866506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.866811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.866834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.867028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.867039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.867232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.867243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.867570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.867580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.867943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.867953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.868230] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.868239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.868537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.868547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.868839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.868849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.869018] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.869029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.869327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.869337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.869657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.869667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.869937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.869947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.870275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.870285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.870484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.870493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.870816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.870826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.871144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.871154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.871436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.871452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.871760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.871770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.872076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.872085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.956 [2024-11-06 09:08:57.872374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.956 [2024-11-06 09:08:57.872386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.956 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.872696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.872706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.873007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.873017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.873399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.873409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.873708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.873718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.874046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.874058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.874172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.874181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.874491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.874501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.874784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.874795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.875104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.875113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.875349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.875359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.875666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.875676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.875871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.875882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.876204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.876214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.876530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.876540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.876809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.876819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.877108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.877117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.877415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.877425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.877720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.877730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.878010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.878020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.878309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.878318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.878637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.878647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.878958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.878969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.879278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.879287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.879569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.879579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.879872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.879882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.880228] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.880239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.880532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.880541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.880752] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.880763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.881083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.881093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.881270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.881281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.881551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.881561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.881919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.957 [2024-11-06 09:08:57.881929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.957 qpair failed and we were unable to recover it. 00:33:07.957 [2024-11-06 09:08:57.882237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.882247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.882543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.882553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.882769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.882780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.883122] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.883132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.883439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.883448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.883780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.883790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.884088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.884098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.884383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.884392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.884683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.884696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.885000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.885010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.885294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.885304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.885498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.885507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.885838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.885848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.886138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.886148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.886432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.886443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.886782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.886792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.886975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.886986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.887172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.887182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.887483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.887492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.887780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.887790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.888113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.888123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.888428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.888438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.888808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.888818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.889012] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.889022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.889347] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.889356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.889639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.889649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.889919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.889929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.890261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.890272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.890575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.890585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.890894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.890904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.891216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.891227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.891507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.891517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.891816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.891826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.892143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.892153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.892421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.892431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.892758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.892770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.893082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.958 [2024-11-06 09:08:57.893092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.958 qpair failed and we were unable to recover it. 00:33:07.958 [2024-11-06 09:08:57.893406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.893416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.893700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.893711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.894041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.894051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.894353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.894363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.894679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.894690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.894992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.895003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.895345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.895356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.895577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.895587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.895899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.895910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.896219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.896228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.896530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.896539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.896842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.896852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.897169] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.897179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.897490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.897499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.897616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.897626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.897926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.897936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.898253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.898264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.898672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.898682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.898987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.898997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.899300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.899310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.899625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.899635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.899963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.899973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.900349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.900359] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.900664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.900674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.900991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.901002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.901304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.901314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.901655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.901665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.901979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.959 [2024-11-06 09:08:57.901990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.959 qpair failed and we were unable to recover it. 00:33:07.959 [2024-11-06 09:08:57.902322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.902333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.902658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.902668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.902984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.902995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.903296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.903306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.903614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.903624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.903957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.903968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.904300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.904310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.904613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.904624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.904916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.904927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.905299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.905309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.905588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.905598] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.905791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.905804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.906109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.906119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.906424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.906434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.906774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.906788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.907073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.907083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.907386] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.907396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.907712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.907721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.907832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.907842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.908115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.908125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.908430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.908440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.908729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.908739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.908974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.908984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.909338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.909347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.909664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.909674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.909987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.909998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.910281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.910290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.910608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.910618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.910919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.910930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.911285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.960 [2024-11-06 09:08:57.911295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.960 qpair failed and we were unable to recover it. 00:33:07.960 [2024-11-06 09:08:57.911606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.911615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.911903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.911913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.912228] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.912238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.912413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.912424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.912753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.912763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.913078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.913088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.913394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.913404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.913670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.913681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.913999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.914012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.914346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.914357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.914670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.914680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.914997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.915008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.915313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.915323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.915689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.915699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.916002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.916012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.916325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.916335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.916646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.916655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.916962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.916972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.917303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.917313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.917375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.917385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.917688] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.917699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.917874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.917884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.918192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.918202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.918522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.918532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.918811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.918821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.918988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.918997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.919298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.919308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.961 [2024-11-06 09:08:57.919634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.961 [2024-11-06 09:08:57.919644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.961 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.919930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.919940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.920241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.920253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.920552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.920562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.920859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.920870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.921242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.921252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.921547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.921558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.921751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.921762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.921959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.921969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.922173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.922183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.922520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.922529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.922813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.922824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.923144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.923155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.923457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.923468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.923636] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.923646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.923918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.923929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.924259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.924269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.924428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.924440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.924760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.924770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.925093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.925104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.925415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.925425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.925729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.925739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.926058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.926071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.926449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.926459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.926790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.926801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.927102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.927112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.927399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.927409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.927730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.927739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.928044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.928055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.962 [2024-11-06 09:08:57.928240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.962 [2024-11-06 09:08:57.928251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.962 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.928550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.928560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.928755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.928773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.929054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.929064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.929340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.929349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.929686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.929695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.930067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.930078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.930406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.930417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.930723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.930733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.931050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.931060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.931363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.931373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.931685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.931695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.932010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.932021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.932361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.932372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.932674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.932684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.932998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.933009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.933217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.933228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.933535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.933545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.933865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.933875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.934175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.934185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.934495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.934505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.934883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.934893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.935210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.935219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.935531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.935540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.935845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.935855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.936239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.963 [2024-11-06 09:08:57.936249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.963 qpair failed and we were unable to recover it. 00:33:07.963 [2024-11-06 09:08:57.936517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.936527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.936851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.936861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.937049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.937059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.937382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.937392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.937705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.937715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.938027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.938037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.938324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.938339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.938637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.938648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.938851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.938861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.939195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.939205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.939592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.939602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.939902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.939913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.940235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.940245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.940537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.940548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.940855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.940865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.941188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.941199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.941503] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.941513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.941808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.941818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.942138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.942148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.942435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.942446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.942751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.942761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.943031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.943041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.943357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.943367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.943659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.943669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.943987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.943998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.944321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.944332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.944639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.944648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.944944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.944954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.945268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.945278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.945538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.945548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.945768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.945779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.946090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.946100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.946439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.964 [2024-11-06 09:08:57.946449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.964 qpair failed and we were unable to recover it. 00:33:07.964 [2024-11-06 09:08:57.946769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.946780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.947100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.947111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.947322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.947335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.947650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.947659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.947980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.947991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.948299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.948309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.948590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.948600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.948915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.948926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.949131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.949141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.949456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.949465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.949750] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.949760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.950076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.950086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.950320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.950330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.950642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.950652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.950942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.950952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.951270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.951280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.951564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.951580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.951907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.951919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.952238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.952248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.952430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.952440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.952751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.952762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.953116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.953126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.953399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.953411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.953711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.953722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.954051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.954062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.954297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.954307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.954616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.954627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.954861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.965 [2024-11-06 09:08:57.954873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.965 qpair failed and we were unable to recover it. 00:33:07.965 [2024-11-06 09:08:57.955208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.955218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.955532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.955542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.955936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.955947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.956262] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.956272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.956490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.956501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.956723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.956733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.957035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.957046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.957355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.957365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.957753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.957763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.958089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.958099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.958378] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.958388] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.958690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.958700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.958982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.958992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.959281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.959292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.959569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.959579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.959883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.959894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.960213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.960224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.960509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.960520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.960795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.960806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.961119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.961128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.961426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.961436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.961749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.961760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.962073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.962084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.962386] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.962396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.962692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.962702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.963025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.963035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.963315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.963325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.963641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.963651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.966 [2024-11-06 09:08:57.963873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.966 [2024-11-06 09:08:57.963885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.966 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.964239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.964249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.964548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.964557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.964849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.964859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.965138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.965148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.965433] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.965443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.965769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.965780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.966079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.966090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.966392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.966403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.966715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.966725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.967041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.967051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.967359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.967369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.967685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.967696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.968065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.968076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.968409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.968422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.968736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.968758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.969044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.969054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.969336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.969353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.969680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.969691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.970002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.970012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.970192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.970203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.970524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.970534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.970756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.970767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.971143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.971154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.971355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.971366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.971652] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.971662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.971977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.971988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.972285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.972295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.972601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.972612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.972912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.972923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.973220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.973230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.973546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.973555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.973862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.973872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.974182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.974193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.974573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.974584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.974897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.967 [2024-11-06 09:08:57.974907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.967 qpair failed and we were unable to recover it. 00:33:07.967 [2024-11-06 09:08:57.975204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.975214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.975395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.975405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.975700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.975711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.976017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.976027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.976388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.976398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.976683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.976693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.977009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.977021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.977320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.977331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.977670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.977680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.977998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.978009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.978194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.978204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.978549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.978560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.978856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.978866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.979200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.979211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.979495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.979505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.979812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.979822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.980139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.980149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.980461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.980471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.980660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.980671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.980998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.981011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.981305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.981315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.981694] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.981704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.982008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.982020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.982323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.982332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.982630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.982640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.982959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.982969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.983264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.983275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.983577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.983586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.983900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.983910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.984228] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.984237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.984555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.984565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.984872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.968 [2024-11-06 09:08:57.984882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.968 qpair failed and we were unable to recover it. 00:33:07.968 [2024-11-06 09:08:57.985180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.985191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.985509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.985519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.985896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.985906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.986216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.986226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.986449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.986459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.986779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.986789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.987086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.987096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.987409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.987419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.987705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.987716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.987864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.987876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.988171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.988181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.988565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.988575] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.988887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.988897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.989198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.989207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.989524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.989536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.989721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.989732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.990016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.990027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.990335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.990345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.990516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.990525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.990875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.990885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.991074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.991083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.991351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.991360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.969 [2024-11-06 09:08:57.991530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.969 [2024-11-06 09:08:57.991540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.969 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.991833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.991843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.992180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.992190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.992471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.992489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.992810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.992821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.993146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.993156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.993486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.993497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.993802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.993812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.994119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.994128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.994471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.994481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.994763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.994773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.995088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.995097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.995258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.995269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.995494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.995504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.995808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.995818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.996131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.996141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.996329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.996338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.996720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.996730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.996902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.996914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.997259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.997268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.997604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.997614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.997914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.997924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.998173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.998184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.998344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.998355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.998592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.998603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.998855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.998865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.999159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.999169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.999338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.999347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.970 [2024-11-06 09:08:57.999569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.970 [2024-11-06 09:08:57.999581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.970 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:57.999878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:57.999889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.000201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.000211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.000420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.000430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.000682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.000693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.001016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.001028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.001312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.001322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.001479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.001491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.001805] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.001815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.002096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.002107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.002415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.002425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.002733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.002742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.003058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.003067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.003378] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.003387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.003703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.003714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.003924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.003935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.004149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.004159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.004472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.004482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.004788] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.004799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.005110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.005121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.005409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.005419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.005706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.005716] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.006034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.006045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.006329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.006339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.006629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.006639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.006965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.006975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.007284] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.007295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.007477] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.007487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.007822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.007833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.971 [2024-11-06 09:08:58.008119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.971 [2024-11-06 09:08:58.008129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.971 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.008357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.008367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.008676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.008686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.009011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.009024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.009356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.009366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.009568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.009578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.009786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.009797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.010030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.010040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.010358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.010368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.010685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.010696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.010898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.010908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.011237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.011247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.011545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.011555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.011891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.011902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.012224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.012234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.012561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.012572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.012856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.012866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Write completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Write completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Write completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Write completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Write completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Write completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 Read completed with error (sct=0, sc=8) 00:33:07.972 starting I/O failed 00:33:07.972 [2024-11-06 09:08:58.013162] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:33:07.972 [2024-11-06 09:08:58.013413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.013439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.013754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.013774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.014013] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.972 [2024-11-06 09:08:58.014026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.972 qpair failed and we were unable to recover it. 00:33:07.972 [2024-11-06 09:08:58.014372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.014385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.014714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.014725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.015130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.015149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.015460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.015474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.018757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.018786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.019166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.019183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.019494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.019507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.019827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.019839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.020174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.020189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.020510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.020524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.020858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.020872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.021202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.021213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.021522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.021536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.021849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.021863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.022194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.022207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.022509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.022520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.022839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.022854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.023064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.023078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.023377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.023389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.023683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.023695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.023899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.023914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.024245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.024259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.024643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.024656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.024978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.024991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.025182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.025195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.025531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.025545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.025895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.025910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.026125] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.026137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.026266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.026276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.026721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.026774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.027195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.027237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.027551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.027564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.973 qpair failed and we were unable to recover it. 00:33:07.973 [2024-11-06 09:08:58.027739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.973 [2024-11-06 09:08:58.027757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.974 qpair failed and we were unable to recover it. 00:33:07.974 [2024-11-06 09:08:58.028186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.974 [2024-11-06 09:08:58.028223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.974 qpair failed and we were unable to recover it. 00:33:07.974 [2024-11-06 09:08:58.028574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.974 [2024-11-06 09:08:58.028587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.974 qpair failed and we were unable to recover it. 00:33:07.974 [2024-11-06 09:08:58.028975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.974 [2024-11-06 09:08:58.029013] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.974 qpair failed and we were unable to recover it. 00:33:07.974 [2024-11-06 09:08:58.029329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.974 [2024-11-06 09:08:58.029342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.974 qpair failed and we were unable to recover it. 00:33:07.974 [2024-11-06 09:08:58.029651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.974 [2024-11-06 09:08:58.029661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.974 qpair failed and we were unable to recover it. 00:33:07.974 [2024-11-06 09:08:58.029964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.974 [2024-11-06 09:08:58.029975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.974 qpair failed and we were unable to recover it. 00:33:07.974 [2024-11-06 09:08:58.030258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:07.974 [2024-11-06 09:08:58.030268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:07.974 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.030602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.030615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.030835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.030847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.031220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.031231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.031519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.031530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.031875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.031887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.032201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.032212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.032402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.032412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.032724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.032734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.033087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.033114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.033423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.033436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.033743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.033767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.034158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.034173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.034398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.034412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.034653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.034664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.034977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.034989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.035303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.035314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.035706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.035724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.035954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.035970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.036293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.036305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.036628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.036640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.036963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.036980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.037280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.037293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.037673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.037685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.038013] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.038027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.038350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.038363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.038548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.038561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.038868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.038881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.039089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.039103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.039414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.039428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.039751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.253 [2024-11-06 09:08:58.039765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.253 qpair failed and we were unable to recover it. 00:33:08.253 [2024-11-06 09:08:58.040073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.040085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.040415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.040427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.040741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.040762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.041181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.041198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.041504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.041515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.041808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.041820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.042121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.042137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.042325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.042339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.042692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.042710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.044755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.044781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.045090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.045104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.045301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.045315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.045498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.045511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.045801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.045814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.046173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.046184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.046530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.046546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.046755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.046768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.047188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.047201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.047512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.047525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.047845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.047861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.048228] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.048243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.048570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.048583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.048891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.048904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.049203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.049220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.049626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.049641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.049955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.049968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.050332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.050347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.050527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.050546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.050962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.050976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.051311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.051324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.051506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.051519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.051857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.051875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.052199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.052212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.052550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.052562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.052907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.052923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.053249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.053265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.053593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.053607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.053920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.254 [2024-11-06 09:08:58.053933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.254 qpair failed and we were unable to recover it. 00:33:08.254 [2024-11-06 09:08:58.054273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.054288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.056755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.056781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.057103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.057116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.057470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.057488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.057791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.057805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.058146] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.058159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.058501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.058516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.058852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.058867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.059167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.059180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.059501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.059513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.059881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.059897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.060246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.060261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.060619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.060634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.060972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.060984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.061207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.061221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.061551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.061566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.061895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.061909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.062254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.062266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.062622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.062639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.063002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.063017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.063240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.063252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.063602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.063614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.063964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.063982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.064312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.064326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.064525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.064537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.064862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.064874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.065245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.065261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.065445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.065458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.065784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.065805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.066212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.066252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.066547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.066560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.067006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.067041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.067220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.067232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.067571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.067581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.068014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.068050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.068251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.068263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.068560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.068570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.068895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.068907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.255 qpair failed and we were unable to recover it. 00:33:08.255 [2024-11-06 09:08:58.069218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.255 [2024-11-06 09:08:58.069232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.069530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.069540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.069722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.069732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.069994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.070005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.070317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.070328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.070611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.070621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.071001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.071011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.071206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.071217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.071523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.071534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.071859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.071870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.072186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.072197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.072497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.072507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.072802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.072812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.073167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.073176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.073460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.073470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.073775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.073785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.074102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.074112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.074324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.074334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.074650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.074660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.074948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.074958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.075287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.075297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.075680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.075690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.075933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.075943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.076259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.076268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.076634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.076643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.076809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.076821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.077186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.077196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.077529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.077540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.077719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.077730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.078112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.078123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.078289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.078300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.078587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.078600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.078768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.078780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.079000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.079010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.079295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.079305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.079597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.079606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.079858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.079868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.080192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.256 [2024-11-06 09:08:58.080202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.256 qpair failed and we were unable to recover it. 00:33:08.256 [2024-11-06 09:08:58.080485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.080495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.080772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.080782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.081007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.081018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.081326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.081336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.081639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.081648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.081850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.081861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.082186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.082196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.082362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.082373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.082581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.082591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.082926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.082936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.083254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.083263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.083459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.083468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.083727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.083737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.084027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.084037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.084171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.084181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.084448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.084458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.084730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.084741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.085045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.085055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.085438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.085448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.085783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.085794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.086029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.086040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.086341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.086350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.086677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.086686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.087008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.087019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.087320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.087330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.087493] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.087505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.087808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.087818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.087994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.088005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.088377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.088386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.088730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.088740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.089047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.089057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.089402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.257 [2024-11-06 09:08:58.089413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.257 qpair failed and we were unable to recover it. 00:33:08.257 [2024-11-06 09:08:58.089760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.089770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.090072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.090087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.090258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.090269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.090584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.090594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.090901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.090911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.091317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.091326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.091603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.091616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.091816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.091829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.092206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.092218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.092462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.092474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.092800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.092813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.093111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.093123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.093384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.093396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.093693] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.093705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.094118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.094131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.094468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.094481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.094798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.094811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.095130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.095143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.095459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.095471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.095815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.095828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.096140] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.096152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.096433] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.096446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.096642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.096655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.096956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.096968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.097285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.097297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.097602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.097614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.097925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.097938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.098124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.098138] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.098451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.098463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.098789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.098803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.099119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.099131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.099484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.099496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.099809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.099822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.100199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.100211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.100496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.100508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.100793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.100806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.101120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.101132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.101431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.258 [2024-11-06 09:08:58.101443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.258 qpair failed and we were unable to recover it. 00:33:08.258 [2024-11-06 09:08:58.101631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.101643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.101988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.102001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.102285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.102297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.102676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.102691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.102991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.103004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.103329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.103341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.103628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.103640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.104016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.104029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.104361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.104379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.104720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.104737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.105047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.105064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.105459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.105475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.105791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.105809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.106135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.106151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.106464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.106480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.106828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.106846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.107180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.107197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.107504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.107521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.107846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.107864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.108188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.108204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.108547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.108563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.108875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.108892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.109210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.109227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.109532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.109549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.109876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.109893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.110191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.110207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.110522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.110539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.110873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.110890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.111190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.111206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.111404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.111423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.111762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.111780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.112094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.112111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.112443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.112459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.112757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.112775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.113100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.113117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.113484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.113500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.113817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.113835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.114178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.114195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.259 [2024-11-06 09:08:58.114480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.259 [2024-11-06 09:08:58.114496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.259 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.114829] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.114846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.115139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.115155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.115438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.115455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.115853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.115871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.116199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.116219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.116537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.116554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.116885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.116907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.117252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.117273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.117622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.117643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.117964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.117986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.118224] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.118247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.118584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.118605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.118911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.118933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.119233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.119254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.119562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.119583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.119878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.119900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.120136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.120157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.120492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.120512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.120815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.120836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.121168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.121189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.121517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.121538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.121842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.121865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.122198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.122219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.122440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.122461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.122810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.122831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.123147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.123167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.123369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.123392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.123737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.123766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.124085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.124107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.124411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.124432] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.124736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.124763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.125076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.125098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.125407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.125428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.125647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.125669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.125995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.126017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.126398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.126419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.126765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.126787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.127090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.260 [2024-11-06 09:08:58.127111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.260 qpair failed and we were unable to recover it. 00:33:08.260 [2024-11-06 09:08:58.127410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.127431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.127766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.127788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.128107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.128128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.128435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.128455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.128794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.128816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.129155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.129176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.129514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.129539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.129890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.129913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.130217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.130238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.130557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.130586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.130933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.130963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.131307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.131335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.131675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.131704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.131956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.131986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.132369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.132397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.132756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.132785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.133130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.133159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.133496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.133525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.133801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.133831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.134179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.134208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.134527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.134556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.134883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.134914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.135221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.135250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.135475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.135507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.135851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.135881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.136010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.136041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.136303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.136332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.136619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.136648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.137002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.137031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.137380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.137409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.137768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.137798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.138150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.138178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.138517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.138546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.138737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.138783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.139197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.139226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.139486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.139514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.139911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.139943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.140263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.261 [2024-11-06 09:08:58.140292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.261 qpair failed and we were unable to recover it. 00:33:08.261 [2024-11-06 09:08:58.140554] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.140582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.140830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.140862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.141179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.141208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.141535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.141563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.141924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.141955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.142314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.142343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.142706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.142734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.143075] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.143105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.143450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.143486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.143831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.143861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.144222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.144250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.144588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.144618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.144974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.145003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.145396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.145425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.145769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.145801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.146181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.146211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.146553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.146582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.146868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.146898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.147259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.147288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.147522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.147554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.147789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.147822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.148052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.148081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.148412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.148441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.148613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.148644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.149014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.149044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.149363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.149391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.149736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.149783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.150129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.150158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.150505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.150533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.150878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.150908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.151130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.151162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.151525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.151554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.151914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.151944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.262 [2024-11-06 09:08:58.152294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.262 [2024-11-06 09:08:58.152322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.262 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.152582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.152610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.152895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.152927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.153251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.153280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.153625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.153654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.154037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.154067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.154365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.154394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.154634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.154664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.154997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.155026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.155375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.155404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.155731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.155770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.156094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.156122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.156462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.156491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.156726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.156766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.157161] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.157189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.157526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.157554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.157901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.157932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.158295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.158324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.158675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.158704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.159053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.159083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.159425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.159454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.159793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.159824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.160220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.160249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.160596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.160624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.160965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.160995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.161229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.161260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.161606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.161635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.161892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.161921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.162258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.162287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.162534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.162565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.162920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.162950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.163316] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.163345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.163688] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.163717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.164054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.164083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.164404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.164433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.164772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.164802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.165140] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.165168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.165506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.165536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.165862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.263 [2024-11-06 09:08:58.165892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.263 qpair failed and we were unable to recover it. 00:33:08.263 [2024-11-06 09:08:58.166258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.166287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.166632] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.166660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.167007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.167037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.167380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.167415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.167729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.167765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.168114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.168143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.168488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.168516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.168931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.168960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.169223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.169252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.169674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.169703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.170111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.170142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.170486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.170515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.170863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.170894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.171240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.171269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.171610] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.171638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.171989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.172020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.172368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.172397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.172753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.172784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.173038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.173067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.173428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.173456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.173829] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.173860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.174204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.174232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.174569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.174597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.174945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.174975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.175333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.175361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.175722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.175759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.176129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.176158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.176498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.176526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.176764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.176794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.177042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.177071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.177423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.177452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.177801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.177831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.178178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.178206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.178531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.178559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.178964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.178993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.179342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.179371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.179717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.179753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.180096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.264 [2024-11-06 09:08:58.180125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.264 qpair failed and we were unable to recover it. 00:33:08.264 [2024-11-06 09:08:58.180468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.180496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.180769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.180799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.181190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.181219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.181542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.181570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.181909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.181940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.182309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.182344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.182682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.182710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.183057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.183087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.183449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.183477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.183684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.183715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.184126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.184156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.184459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.184487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.184816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.184846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.185197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.185226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.185566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.185594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.185938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.185968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.186320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.186349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.186697] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.186725] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.187070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.187100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.187343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.187372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.187734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.187770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.188036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.188065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.188412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.188441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.188767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.188797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.189152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.189180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.189526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.189554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.189897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.189927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.190286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.190315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.190575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.190604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.190969] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.190999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.191348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.191377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.191717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.191745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.192097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.192127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.192374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.192402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.192821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.192852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.193097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.193129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.193506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.193534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.193877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.193908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.265 qpair failed and we were unable to recover it. 00:33:08.265 [2024-11-06 09:08:58.194275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.265 [2024-11-06 09:08:58.194304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.194652] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.194680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.195014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.195045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.195333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.195361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.195690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.195718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.196096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.196125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.196468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.196496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.196845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.196881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.197229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.197258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.197600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.197629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.197877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.197911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.198235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.198264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.198606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.198634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.198988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.199017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.199373] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.199401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.199753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.199784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.200170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.200198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.200536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.200564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.200942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.200972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.201324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.201352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.201702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.201731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.202086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.202115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.202460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.202489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.202832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.202861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.203205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.203233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.203578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.203607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.203948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.203978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.204309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.204337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.204706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.204735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.205007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.205037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.205390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.205419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.205784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.205816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.206134] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.206163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.206509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.206537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.266 qpair failed and we were unable to recover it. 00:33:08.266 [2024-11-06 09:08:58.206781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.266 [2024-11-06 09:08:58.206814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.207105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.207134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.207458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.207486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.207827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.207857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.208198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.208226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.208569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.208597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.208934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.208964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.209320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.209349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.209706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.209735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.210087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.210117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.210358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.210387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.210744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.210787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.211138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.211167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.211522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.211557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.211926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.211956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.212310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.212338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.212684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.212712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.213058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.213088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.213436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.213464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.213811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.213842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.214099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.214127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.214538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.214567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.214921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.214951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.215317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.215346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.215590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.215618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.215924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.215954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.216214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.216244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.216582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.216610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.217004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.217034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.217375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.217404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.217763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.217793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.218138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.218167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.218516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.218545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.218956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.218985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.219333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.219361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.219802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.219832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.220173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.220202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.220516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.220544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.267 [2024-11-06 09:08:58.220792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.267 [2024-11-06 09:08:58.220821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.267 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.221065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.221096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.221456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.221485] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.221819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.221850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.222187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.222216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.222564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.222593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.222926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.222955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.223262] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.223291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.223428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.223456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.223823] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.223853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.224226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.224254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.224674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.224702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.225089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.225118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.225463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.225491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.225836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.225866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.226221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.226257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.226605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.226634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.226996] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.227026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.227364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.227393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.227767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.227797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.228154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.228183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.228540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.228569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.228948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.228978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.229301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.229330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.229564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.229593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.229966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.229997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.230343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.230372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.230715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.230743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.231100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.231129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.231473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.231501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.231838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.231868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.232212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.232240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.232581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.232608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.232954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.232984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.233295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.233324] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.233669] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.233697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.234099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.234129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.234460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.234489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.234836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.268 [2024-11-06 09:08:58.234865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.268 qpair failed and we were unable to recover it. 00:33:08.268 [2024-11-06 09:08:58.235213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.235242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.235574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.235602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.235956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.235986] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.236338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.236368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.236718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.236753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.237145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.237175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.237422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.237454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.237801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.237833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.238079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.238108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.238530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.238559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.238970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.238999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.239352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.239380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.239704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.239732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.240155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.240185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.240528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.240557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.240809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.240838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.241227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.241263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.241492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.241521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.241674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.241707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.242056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.242085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.242305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.242337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.242692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.242720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.243055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.243084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.243409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.243437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.243764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.243793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.244148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.244177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.244414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.244442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.244673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.244705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.245057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.245087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.245329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.245357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.245600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.245630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.245896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.245928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.246271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.246299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.246526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.246558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.246896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.246926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.247244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.247272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.247628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.247657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.248038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.248068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.269 qpair failed and we were unable to recover it. 00:33:08.269 [2024-11-06 09:08:58.248311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.269 [2024-11-06 09:08:58.248339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.248694] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.248723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.248972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.249002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.249244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.249272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.249600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.249628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.249939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.249970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.250319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.250348] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.250524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.250556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.250899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.250929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.251354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.251383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.251718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.251755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.251995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.252024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.252242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.252271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.252509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.252538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.252904] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.252934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.253301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.253330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.253573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.253605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.253971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.254002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.254345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.254380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.254715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.254743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.255097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.255126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.255463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.255492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.255844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.255874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.256198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.256226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.256563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.256591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.256973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.257002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.257348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.257376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.257720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.257770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.258121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.258150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.258497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.258526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.258877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.258908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.259270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.259298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.259650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.259679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.259924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.259954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.260312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.260340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.260699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.260728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.261067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.261096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.261437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.261466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.270 qpair failed and we were unable to recover it. 00:33:08.270 [2024-11-06 09:08:58.261706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.270 [2024-11-06 09:08:58.261735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.262097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.262126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.262474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.262502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.262853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.262884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.263199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.263227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.263575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.263604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.263960] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.263990] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.264376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.264405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.264756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.264786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.265132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.265161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.265493] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.265521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.265864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.265894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.266235] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.266263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.266605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.266633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.266975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.267005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.267350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.267379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.267722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.267758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.268022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.268051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.268402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.268431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.268777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.268807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.269133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.269167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.269502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.269531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.269882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.269912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.270212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.270240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.270576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.270605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.270936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.270966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.271288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.271317] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.271691] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.271719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.272061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.272090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.272432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.272460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.272808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.272839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.273192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.273220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.273556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.271 [2024-11-06 09:08:58.273584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.271 qpair failed and we were unable to recover it. 00:33:08.271 [2024-11-06 09:08:58.273927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.273956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.274302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.274332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.274680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.274708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.275053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.275082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.275307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.275335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.275692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.275721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.276096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.276124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.276376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.276405] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.276762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.276792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.277139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.277168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.277505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.277534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.277788] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.277818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.278203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.278231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.278554] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.278583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.278966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.278998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.279348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.279377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.279706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.279734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.280077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.280106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.280448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.280477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.280826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.280855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.281198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.281226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.281477] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.281506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.281877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.281906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.282249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.282277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.282616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.282646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.282989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.283019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.283362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.283391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.283731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.283778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.284014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.284045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.284384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.284412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.284757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.284788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.285133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.285162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.285505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.285534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.285887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.285917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.286248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.286276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.286579] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.286607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.286899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.286929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.287297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.287325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.272 qpair failed and we were unable to recover it. 00:33:08.272 [2024-11-06 09:08:58.287673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.272 [2024-11-06 09:08:58.287702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.288041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.288070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.288418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.288446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.288808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.288839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.289204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.289232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.289576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.289603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.289957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.289987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.290331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.290360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.290706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.290734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.291080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.291109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.291450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.291480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.291767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.291797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.292150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.292178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.292524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.292553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.292934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.292964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.293280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.293308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.293649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.293678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.294078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.294109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.294458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.294487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.294833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.294863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.295228] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.295256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.295605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.295634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.296019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.296048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.296395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.296423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.296763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.296793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.297142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.297171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.297519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.297549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.297893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.297922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.298246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.298275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.298628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.298663] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.298988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.299019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.299362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.299392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.299736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.299773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.300106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.300135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.300486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.300515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.300753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.300783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.301171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.301200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.301522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.301551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.273 [2024-11-06 09:08:58.301899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.273 [2024-11-06 09:08:58.301929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.273 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.302293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.302321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.302669] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.302697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.302986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.303016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.303363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.303391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.303769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.303800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.304166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.304195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.304540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.304568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.304822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.304852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.305214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.305243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.305593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.305622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.305963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.305992] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.306333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.306361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.306706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.306734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.307084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.307113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.307455] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.307483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.307725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.307762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.308113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.308141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.308496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.308527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.308912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.308942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.309339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.309368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.309764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.309795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.310159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.310187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.310545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.310574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.310827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.310857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.311263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.311292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.311639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.311668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.312007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.312038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.312379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.312408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.312757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.312787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.313168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.313196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.313543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.313578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.313927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.313958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.314292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.314321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.314685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.314714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.315064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.315095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.315455] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.315484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.315834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.315864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.316217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.274 [2024-11-06 09:08:58.316246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.274 qpair failed and we were unable to recover it. 00:33:08.274 [2024-11-06 09:08:58.316590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.316618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.317016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.317045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.317392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.317420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.317795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.317825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.318184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.318213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.318563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.318591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.318961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.318991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.319333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.319362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.319722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.319761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.320117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.320146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.320466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.320494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.320845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.320875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.321216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.321245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.321589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.321618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.321971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.322002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.322355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.322384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.322737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.322775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.323101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.323130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.323476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.323504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.323763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.323796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.324133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.324162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.324506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.324535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.324806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.324835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.325195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.325224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.325566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.325595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.325954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.325984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.326324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.326353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.326661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.326690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.327047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.327077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.327418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.327446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.327790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.327820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.328177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.328205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.328554] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.328589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.328870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.328900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.275 qpair failed and we were unable to recover it. 00:33:08.275 [2024-11-06 09:08:58.329262] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.275 [2024-11-06 09:08:58.329292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.329637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.329666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.330017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.330048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.330390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.330419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.330769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.330800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.331149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.331178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.331429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.331458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.331815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.331845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.332181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.332209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.332548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.332577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.332912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.332941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.333191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.333220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.333544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.333573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.333929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.333959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.334317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.334347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.334703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.334733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.335117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.335146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.335490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.335519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.335861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.335891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.336241] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.336269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.336623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.336651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.337023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.337053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.337396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.337426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.337764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.337794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.338139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.338167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.338522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.338552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.338902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.338932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.339323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.339351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.339699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.339728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.340068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.340098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.340449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.340477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.340810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.340840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.341187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.341217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.341575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.341603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.341992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.342022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.342364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.342392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.342741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.342781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.343073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.343100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.276 qpair failed and we were unable to recover it. 00:33:08.276 [2024-11-06 09:08:58.343449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.276 [2024-11-06 09:08:58.343483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.277 qpair failed and we were unable to recover it. 00:33:08.277 [2024-11-06 09:08:58.343826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.277 [2024-11-06 09:08:58.343856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.277 qpair failed and we were unable to recover it. 00:33:08.277 [2024-11-06 09:08:58.344240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.277 [2024-11-06 09:08:58.344268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.277 qpair failed and we were unable to recover it. 00:33:08.277 [2024-11-06 09:08:58.344627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.277 [2024-11-06 09:08:58.344655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.277 qpair failed and we were unable to recover it. 00:33:08.277 [2024-11-06 09:08:58.345004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.277 [2024-11-06 09:08:58.345034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.277 qpair failed and we were unable to recover it. 00:33:08.277 [2024-11-06 09:08:58.345382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.277 [2024-11-06 09:08:58.345411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.277 qpair failed and we were unable to recover it. 00:33:08.277 [2024-11-06 09:08:58.345768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.277 [2024-11-06 09:08:58.345797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.277 qpair failed and we were unable to recover it. 00:33:08.277 [2024-11-06 09:08:58.346131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.277 [2024-11-06 09:08:58.346160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.277 qpair failed and we were unable to recover it. 00:33:08.277 [2024-11-06 09:08:58.346407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.277 [2024-11-06 09:08:58.346435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.277 qpair failed and we were unable to recover it. 00:33:08.277 [2024-11-06 09:08:58.346852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.277 [2024-11-06 09:08:58.346881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.277 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.347246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.347277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.347623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.347652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.348000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.348029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.348448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.348477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.348800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.348830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.349082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.349111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.349440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.349468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.349814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.349844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.350203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.350232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.350571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.350599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.350949] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.350979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.351332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.351360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.351706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.351734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.352093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.352122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.352480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.352508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.352781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.352810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.353175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.353204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.353549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.353583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.353929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.353960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.354213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.354241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.354600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.354629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.355009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.355038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.355388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.355416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.355666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.355694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.356069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.356098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.356443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.356470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.356820] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.356850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.357188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.357216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.357565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.357592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.358023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.358053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.358273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.553 [2024-11-06 09:08:58.358302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.553 qpair failed and we were unable to recover it. 00:33:08.553 [2024-11-06 09:08:58.358654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.358683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.359043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.359073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.359418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.359447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.359874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.359904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.360244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.360272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.360593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.360622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.360968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.360998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.361255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.361284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.361666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.361695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.362112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.362142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.362466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.362494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.362813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.362843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.363181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.363209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.363572] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.363600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.364024] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.364054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.364392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.364421] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.364778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.364807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.365153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.365182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.365534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.365562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.365903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.365934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.366280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.366309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.366661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.366689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.367030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.367061] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.367383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.367412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.367735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.367772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.368102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.368131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.368473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.368507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.368836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.368866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.369225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.369254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.369599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.369627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.370027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.370058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.370414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.370443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.370683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.370712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.370976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.371009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.371336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.371365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.371721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.371759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.371983] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.372015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.554 [2024-11-06 09:08:58.372367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.554 [2024-11-06 09:08:58.372396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.554 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.372760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.372791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.373145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.373173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.373528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.373557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.373885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.373916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.374137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.374165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.374609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.374638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.375014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.375044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.375383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.375412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.375656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.375685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.376043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.376073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.376300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.376331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.376589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.376618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.376943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.376974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.377325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.377354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.377713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.377742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.378087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.378117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.378459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.378487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.378716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.378757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.379104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.379134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.379482] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.379510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.379844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.379874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.380242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.380271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.380523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.380551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.380769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.380801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.381167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.381197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.381541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.381569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.381954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.381984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.382346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.382375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.382726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.382771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.383119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.383148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.383448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.383476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.383827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.383857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.384209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.384237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.384583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.384612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.385008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.385039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.385385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.385414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.385764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.385794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.386123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.555 [2024-11-06 09:08:58.386152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.555 qpair failed and we were unable to recover it. 00:33:08.555 [2024-11-06 09:08:58.386492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.386521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.386887] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.386917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.387278] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.387307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.387640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.387668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.388000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.388030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.388381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.388410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.388765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.388794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.389151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.389180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.389545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.389574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.389930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.389960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.390325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.390354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.390681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.390709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.391047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.391076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.391429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.391457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.391717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.391753] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.392163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.392193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.392541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.392570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.392922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.392952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.393304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.393333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.393684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.393712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.394096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.394127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.394475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.394504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.394858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.394887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.395232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.395261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.395512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.395540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.395943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.395973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.396316] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.396344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.396691] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.396719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.396989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.397022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.397361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.397390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.397721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.397777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.398017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.398046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.398405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.398433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.398784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.398813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.399229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.399257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.399611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.399640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.400043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.400072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.556 [2024-11-06 09:08:58.400418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.556 [2024-11-06 09:08:58.400446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.556 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.400729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.400768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.401188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.401216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.401565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.401593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.401848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.401879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.402300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.402329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.402671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.402699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.403056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.403086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.403430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.403459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.403788] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.403817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.404041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.404073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.404493] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.404522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.404865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.404894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.405121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.405153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.405506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.405534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.405878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.405908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.406269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.406298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.406641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.406669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.407026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.407056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.407398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.407427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.407778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.407808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.408135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.408163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.408398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.408426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.408758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.408788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.409125] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.409153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.409504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.409532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.409875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.409906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.410253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.410282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.410643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.410671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.411002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.411032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.411375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.411403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.411756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.411786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.412116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.557 [2024-11-06 09:08:58.412144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.557 qpair failed and we were unable to recover it. 00:33:08.557 [2024-11-06 09:08:58.412500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.412534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.412873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.412903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.413267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.413295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.413651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.413679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.414038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.414068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.414413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.414442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.414787] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.414818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.415164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.415192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.415499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.415527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.415877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.415907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.416263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.416292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.416612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.416639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.417015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.417044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.417391] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.417419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.417767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.417798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.418043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.418074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.418431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.418460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.418802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.418833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.419168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.419196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.419542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.419571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.419935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.419965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.420326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.420354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.420698] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.420728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.421078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.421107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.421453] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.421481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.421840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.421870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.422232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.422260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.422606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.422635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.422861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.422893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.423243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.423271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.423632] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.423661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.424018] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.424048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.424410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.424439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.424684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.424715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.425080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.425110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.425359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.425391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.425741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.425787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.426114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.426143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.558 qpair failed and we were unable to recover it. 00:33:08.558 [2024-11-06 09:08:58.426508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.558 [2024-11-06 09:08:58.426536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.426883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.426913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.427128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.427167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.427527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.427557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.427926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.427956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.428186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.428215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.428554] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.428583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.429000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.429029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.429254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.429285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.429529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.429561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.429920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.429950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.430276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.430305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.430649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.430677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.431035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.431066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.431373] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.431401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.431765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.431796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.432160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.432189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.432419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.432448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.432676] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.432707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.433058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.433088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.433480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.433508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.433712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.433741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.434120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.434150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.434478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.434505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.434843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.434873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.435205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.435234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.435563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.435591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.435904] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.435934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.436304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.436333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.436684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.436714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.437067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.437098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.437439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.437467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.437825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.437856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.438082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.438115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.438438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.438466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.438799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.438828] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.439176] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.439205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.439411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.439440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.559 [2024-11-06 09:08:58.439764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.559 [2024-11-06 09:08:58.439794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.559 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.440175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.440204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.440556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.440585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.440928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.440958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.441182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.441221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.441570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.441599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.441969] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.441999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.442331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.442360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.442725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.442765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.443096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.443124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.443505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.443534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.443899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.443930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.444275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.444304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.444659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.444687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.444939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.444972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.445300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.445329] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.445689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.445717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.445926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.445958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.446323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.446352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.446588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.446616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.447004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.447034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.447288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.447316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.447647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.447676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.448041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.448071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.448416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.448445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.448761] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.448790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.449180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.449209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.449558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.449586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.449937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.449967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.450283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.450312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.450651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.450680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.451129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.451160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.451505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.451533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.451884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.451914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.452272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.452301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.452646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.452674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.453052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.453083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.453328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.453357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.453716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.560 [2024-11-06 09:08:58.453745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.560 qpair failed and we were unable to recover it. 00:33:08.560 [2024-11-06 09:08:58.454121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.454150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.454494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.454522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.454875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.454905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.455267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.455296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.455535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.455565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.455859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.455896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.456243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.456271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.456621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.456650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.457063] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.457093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.457454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.457482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.457843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.457874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.458253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.458282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.458630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.458659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.459049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.459079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.459441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.459470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.459793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.459825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.460160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.460190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.460539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.460567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.460918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.460948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.461328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.461357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.461685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.461712] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.462088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.462118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.462365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.462397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.462640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.462669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.463025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.463055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.463474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.463502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.463830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.463860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.464121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.464149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.464490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.464519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.464872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.464903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.465291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.465320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.465715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.465744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.466155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.466185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.466430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.466459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.466796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.466825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.467178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.467207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.467356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.467384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.467799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.467830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.561 [2024-11-06 09:08:58.468164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.561 [2024-11-06 09:08:58.468193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.561 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.468539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.468567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.468909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.468939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.469295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.469323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.469672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.469700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.470098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.470128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.470475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.470503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.470849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.470884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.471217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.471246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.471451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.471480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.471817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.471847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.472180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.472209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.472578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.472607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.472953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.472983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.473315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.473343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.473615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.473644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.473995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.474025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.474396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.474424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.474786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.474816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.475165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.475194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.475555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.475584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.476002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.476032] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.476359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.476387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.476754] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.476784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.477060] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.477088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.477333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.477361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.477592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.477625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.477976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.478006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.478339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.478368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.478729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.478767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.479136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.479164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.479424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.479452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.479707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.479740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.480128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.562 [2024-11-06 09:08:58.480157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.562 qpair failed and we were unable to recover it. 00:33:08.562 [2024-11-06 09:08:58.480470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.480500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.480760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.480791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.481131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.481160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.481376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.481408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.481763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.481794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.482035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.482067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.482396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.482425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.482781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.482811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.483152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.483181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.483533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.483562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.483986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.484017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.484408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.484437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.484659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.484687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.485098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.485135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.485483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.485512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.485868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.485899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.486259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.486288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.486637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.486666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.487017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.487047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.487369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.487397] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.487744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.487783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.488123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.488151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.488508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.488537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.488782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.488815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.489178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.489206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.489549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.489577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.489916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.489946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.490206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.490235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.490645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.490673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.490996] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.491026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.491367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.491396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.491741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.491782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.492119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.492148] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.492503] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.492531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.492879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.492909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.493269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.493297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.493649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.493678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.494026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.563 [2024-11-06 09:08:58.494056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.563 qpair failed and we were unable to recover it. 00:33:08.563 [2024-11-06 09:08:58.494474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.494504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.494854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.494884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.495215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.495245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.495590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.495618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.496041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.496071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.496396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.496424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.496774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.496804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.497163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.497191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.497532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.497560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.497964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.497994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.498346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.498375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.498630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.498658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.499006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.499035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.499388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.499419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.499764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.499794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.500143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.500178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.500529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.500557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.500935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.500966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.501335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.501363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.501708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.501737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.502082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.502112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.502458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.502487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.502842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.502871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.503273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.503302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.503541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.503573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.503773] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.503801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.504118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.504147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.504507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.504536] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.504797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.504826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.505283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.505312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.505665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.505694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.506051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.506081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.506427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.506455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.506875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.506905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.507221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.507250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.507594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.507622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.507963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.507993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.508350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.508379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.564 qpair failed and we were unable to recover it. 00:33:08.564 [2024-11-06 09:08:58.508730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.564 [2024-11-06 09:08:58.508770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.509105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.509133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.509489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.509517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.509868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.509899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.510261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.510290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.510640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.510669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.510908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.510938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.511276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.511305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.511660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.511689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.512053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.512083] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.512390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.512419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.512680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.512708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.513124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.513153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.513507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.513535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.513892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.513922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.514272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.514300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.514595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.514624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.514979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.515015] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.515364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.515392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.515737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.515775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.516107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.516137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.516494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.516522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.516938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.516968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.517362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.517391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.517710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.517739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.518111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.518140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.518484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.518512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.518857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.518888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.519248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.519277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.519620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.519648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.519982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.520011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.520405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.520434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.520682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.520711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.521093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.521125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.521463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.521492] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.521851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.521882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.522139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.522172] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.522516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.522545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.565 [2024-11-06 09:08:58.522892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.565 [2024-11-06 09:08:58.522922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.565 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.523294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.523323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.523674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.523702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.524054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.524084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.524444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.524473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.524846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.524875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.525239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.525268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.525619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.525648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.526006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.526036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.526382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.526410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.526678] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.526706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.527097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.527127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.527470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.527498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.527862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.527893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.528268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.528297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.528646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.528674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.529073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.529102] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.529449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.529478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.529830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.529861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.530223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.530258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.530602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.530630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.530975] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.531006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.531358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.531386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.531762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.531792] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.532136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.532165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.532515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.532543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.532879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.532908] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.533126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.533158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.533411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.533440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.533802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.533832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.534202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.534231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.534577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.534605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.534901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.534931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.535313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.535342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.535690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.566 [2024-11-06 09:08:58.535719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.566 qpair failed and we were unable to recover it. 00:33:08.566 [2024-11-06 09:08:58.536096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.536126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.536476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.536505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.536858] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.536888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.537249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.537277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.537621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.537650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.538002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.538031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.538354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.538383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.538723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.538760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.539097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.539126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.539476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.539505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.539852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.539882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.540112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.540147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.540534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.540562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.540903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.540933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.541299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.541327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.541570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.541601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.541966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.541997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.542325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.542353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.542609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.542638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.542977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.543006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.543362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.543391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.543726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.543763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.544097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.544125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.544474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.544502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.544860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.544890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.545260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.545288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.545633] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.545661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.545998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.546028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.546373] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.546402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.546763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.546793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.547068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.547096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.547446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.547474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.547821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.547851] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.548213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.548243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.548573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.548602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.548939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.548969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.549320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.549350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.549614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.567 [2024-11-06 09:08:58.549642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.567 qpair failed and we were unable to recover it. 00:33:08.567 [2024-11-06 09:08:58.549899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.549930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.550291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.550319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.550671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.550699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.550949] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.550980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.551328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.551356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.551698] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.551726] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.552159] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.552189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.552519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.552548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.552868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.552899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.553233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.553263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.553614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.553643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.553997] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.554027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.554329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.554357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.554687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.554722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.555092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.555122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.555466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.555495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.555846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.555877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.556245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.556273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.556526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.556555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.556893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.556923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.557206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.557234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.557480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.557511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.557880] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.557910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.558259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.558288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.558636] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.558665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.559009] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.559039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.559394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.559423] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.559783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.559813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.560080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.560109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.560452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.560481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.560813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.560843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.561188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.561217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.561569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.561597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.561945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.561974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.562184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.562213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.562551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.562580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.562908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.562937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.568 [2024-11-06 09:08:58.563357] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.568 [2024-11-06 09:08:58.563386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.568 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.563730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.563766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.564099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.564127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.564373] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.564402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.564651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.564679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.565051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.565081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.565425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.565454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.565707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.565735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.566011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.566040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.566314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.566343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.566687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.566715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.567088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.567118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.567473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.567500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.567856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.567886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.568225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.568254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.568612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.568641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.568980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.569016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.569382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.569411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.569764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.569794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.570030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.570059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.570375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.570403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.570765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.570795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.571080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.571108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.571447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.571475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.571802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.571831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.572172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.572201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.572548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.572576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.572942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.572971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.573392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.573420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.573779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.573810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.574189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.574218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.574568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.574597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.574934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.574964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.575231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.575259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.575593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.575621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.576010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.576039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.576402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.576430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.576780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.576811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.577161] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.569 [2024-11-06 09:08:58.577191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.569 qpair failed and we were unable to recover it. 00:33:08.569 [2024-11-06 09:08:58.577521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.577551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.577781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.577811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.578059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.578090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.578452] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.578481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.578852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.578882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.579250] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.579278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.579696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.579724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.580061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.580090] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.580437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.580465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.580814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.580845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.581102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.581134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.581458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.581486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.581852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.581882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.582167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.582195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.582524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.582552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.582899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.582928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.583236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.583265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.583624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.583659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.583888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.583921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.584326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.584354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.584723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.584766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.585078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.585107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.585376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.585404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.585736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.585783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.586126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.586155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.586390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.586419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.586782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.586813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.587163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.587192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.587468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.587496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.587875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.587905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.588273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.588302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.588594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.588623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.589014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.589044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.589395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.589425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.589768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.589798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.590156] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.590185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.590517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.590546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.590895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.590925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.570 [2024-11-06 09:08:58.591275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.570 [2024-11-06 09:08:58.591303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.570 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.591650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.591679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.592020] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.592050] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.592405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.592433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.592780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.592810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.593185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.593213] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.593574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.593603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.593963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.593994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.594345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.594374] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.594719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.594759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.595137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.595168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.595517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.595546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.595893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.595924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.596243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.596272] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.596620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.596648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.596984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.597014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.597352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.597382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.597726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.597763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.598147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.598175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.598540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.598580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.598988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.599018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.599349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.599378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.599721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.599760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.600093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.600121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.600469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.600498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.600872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.600902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.601258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.601286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.601637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.601666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.602008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.602038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.602370] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.602399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.602758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.602789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.603124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.603152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.603500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.603529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.571 [2024-11-06 09:08:58.603881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.571 [2024-11-06 09:08:58.603912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.571 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.604272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.604301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.604553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.604587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.604967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.604997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.605326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.605355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.605702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.605730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.606097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.606126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.606433] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.606462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.606849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.606879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.607234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.607262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.607612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.607641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.608021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.608051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.608401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.608430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.608781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.608811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.609173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.609201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.609558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.609586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.609932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.609963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.610320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.610349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.610722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.610762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.611133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.611162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.611521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.611548] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.611903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.611934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.612285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.612314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.612656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.612685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.613040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.613069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.613420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.613449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.613807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.613843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.614202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.614231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.614580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.614609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.614961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.614991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.615343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.615372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.615720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.615758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.616011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.616040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.616409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.616437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.616678] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.616706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.617109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.617139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.617491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.617520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.617871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.572 [2024-11-06 09:08:58.617903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.572 qpair failed and we were unable to recover it. 00:33:08.572 [2024-11-06 09:08:58.618258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.618287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.618657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.618685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.619041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.619072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.619416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.619444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.619813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.619843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.620193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.620221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.620614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.620643] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.621047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.621077] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.621410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.621438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.621813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.621844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.622221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.622249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.622536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.622565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.622801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.622831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.623212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.623240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.623576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.623605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.623968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.623999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.624339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.624367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.624703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.624732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.625130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.625159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.625502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.625531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.625882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.625913] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.626264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.626293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.626646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.626674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.627050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.627080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.627426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.627454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.627803] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.627833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.628193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.628223] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.628543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.628572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.628886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.628921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.629291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.629321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.629670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.629700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.630113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.630143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.630494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.630523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.630886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.630916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.631256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.631285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.631632] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.631661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.632031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.632060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.632413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.573 [2024-11-06 09:08:58.632442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.573 qpair failed and we were unable to recover it. 00:33:08.573 [2024-11-06 09:08:58.632789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.632819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.633198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.633226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.633578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.633609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.633955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.633985] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.634359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.634389] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.634737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.634775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.635123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.635151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.635504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.635533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.635874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.635905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.636264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.636293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.636643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.636672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.637056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.637086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.637447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.637476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.637848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.637878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.638239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.638268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.638625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.638653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.638993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.639023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.639369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.639399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.639637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.639665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.640018] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.640049] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.640396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.640424] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.640772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.640802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.641056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.641085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.641463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.641491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.641845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.641875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.642220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.642248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.642598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.642626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.643022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.643052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.643407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.643436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.643841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.643872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.644236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.644271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.644623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.644653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.644991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.645021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.645351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.645380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.645740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.645796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.646118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.646147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.646494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.646522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.646857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.646887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.574 [2024-11-06 09:08:58.647240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.574 [2024-11-06 09:08:58.647269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.574 qpair failed and we were unable to recover it. 00:33:08.575 [2024-11-06 09:08:58.647649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.575 [2024-11-06 09:08:58.647678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.575 qpair failed and we were unable to recover it. 00:33:08.575 [2024-11-06 09:08:58.647935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.575 [2024-11-06 09:08:58.647965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.575 qpair failed and we were unable to recover it. 00:33:08.575 [2024-11-06 09:08:58.648329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.575 [2024-11-06 09:08:58.648357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.575 qpair failed and we were unable to recover it. 00:33:08.575 [2024-11-06 09:08:58.648728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.575 [2024-11-06 09:08:58.648765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.575 qpair failed and we were unable to recover it. 00:33:08.575 [2024-11-06 09:08:58.649132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.575 [2024-11-06 09:08:58.649161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.575 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.649523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.649554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.649902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.649935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.650271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.650299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.650646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.650674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.651008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.651038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.651382] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.651411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.651767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.651798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.652092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.652121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.652488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.652517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.652911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.652942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.653300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.653328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.653689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.653717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.654139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.849 [2024-11-06 09:08:58.654169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.849 qpair failed and we were unable to recover it. 00:33:08.849 [2024-11-06 09:08:58.654566] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.654596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.654943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.654973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.655321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.655350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.655704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.655733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.656094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.656123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.656474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.656503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.656864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.656895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.657272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.657301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.657647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.657676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.657919] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.657949] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.658312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.658341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.658691] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.658719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.659017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.659047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.659368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.659403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.659761] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.659791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.660121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.660150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.660489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.660518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.660872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.660902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.661277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.661305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.661661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.661690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.662082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.662113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.662478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.662507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.662860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.662889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.663254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.663282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.663683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.663711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.664066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.664096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.664451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.664480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.664840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.664871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.665127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.665156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.665500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.665528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.665884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.665914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.666265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.666294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.666540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.666568] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.666924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.666953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.667316] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.667344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.667713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.667742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.668147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.668175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.668538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.850 [2024-11-06 09:08:58.668567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.850 qpair failed and we were unable to recover it. 00:33:08.850 [2024-11-06 09:08:58.668817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.668848] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.669275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.669304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.669641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.669670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.670043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.670073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.670422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.670451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.670790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.670820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.671060] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.671091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.671451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.671479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.671829] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.671859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.672251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.672279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.672521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.672553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.672903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.672934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.673295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.673323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.673677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.673706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.674124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.674154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.674504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.674547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.674880] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.674910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.675269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.675297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.675638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.675666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.676025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.676054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.676303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.676335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.676736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.676778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.677200] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.677228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.677580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.677608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.677958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.677989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.678322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.678350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.678704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.678732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.679079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.679108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.679462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.679490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.679841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.679871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.680215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.680245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.680585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.680613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.681018] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.681048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.681400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.681429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.681779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.681808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.682141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.682169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.682517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.682546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.682918] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.682947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.851 [2024-11-06 09:08:58.683280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.851 [2024-11-06 09:08:58.683308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.851 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.683666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.683694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.684057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.684086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.684437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.684465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.684819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.684850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.685248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.685276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.685576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.685604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.685943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.685974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.686303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.686332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.686682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.686711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.687075] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.687105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.687415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.687443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.687810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.687840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.688175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.688203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.688556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.688584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.688921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.688951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.689323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.689353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.689705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.689739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.690025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.690054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.690410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.690438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.690706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.690734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.691114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.691144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.691454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.691482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.691911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.691941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.692288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.692316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.692671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.692699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.693048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.693078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.693421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.693449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.693760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.693791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.694149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.694176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.694528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.694556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.694903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.694934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.695214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.695242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.695491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.695520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.695897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.695927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.696152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.696183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.696551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.696580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.696857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.696888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.852 qpair failed and we were unable to recover it. 00:33:08.852 [2024-11-06 09:08:58.697253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.852 [2024-11-06 09:08:58.697281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.697528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.697557] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.697896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.697926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.698294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.698322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.698666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.698695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.699069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.699099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.699447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.699476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.699852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.699882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.700190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.700219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.700551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.700579] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.700932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.700963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.701213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.701242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.701585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.701615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.702010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.702040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.702437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.702465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.702699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.702731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.703087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.703116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.703448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.703476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.703844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.703874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.704248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.704284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.704608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.704637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.704967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.704997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.705354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.705383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.705734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.705772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.706104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.706132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.706490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.706518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.706888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.706917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.707263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.707292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.707663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.707692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.708051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.708081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.708409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.708438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.708844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.708874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.709248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.709276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.709538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.709566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.709958] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.709987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.710342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.710371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.710719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.710756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.711116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.711145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.853 [2024-11-06 09:08:58.711398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.853 [2024-11-06 09:08:58.711426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.853 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.711799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.711829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.712171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.712199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.712554] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.712582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.712932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.712963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.713216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.713244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.713606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.713634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.713994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.714024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.714356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.714386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.714766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.714796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.715067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.715095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.715437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.715465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.715892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.715922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.716151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.716179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.716435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.716468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.716808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.716838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.717193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.717221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.717464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.717496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.717876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.717906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.718277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.718306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.718670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.718699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.719069] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.719105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.719499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.719528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.719815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.719846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.720207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.720236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.720594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.720622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.720978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.721009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.721350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.721379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.721715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.721743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.722093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.722122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.722472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.722500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.722863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.722892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.854 [2024-11-06 09:08:58.723245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.854 [2024-11-06 09:08:58.723273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.854 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.723627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.723655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.724042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.724073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.724460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.724489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.724837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.724867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.725236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.725264] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.725620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.725648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.725989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.726018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.726379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.726408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.726765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.726796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.727133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.727161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.727514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.727542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.727922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.727953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.728332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.728360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.728712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.728740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.729121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.729151] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.729499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.729534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.729867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.729897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.730250] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.730279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.730628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.730656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.731039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.731068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.731435] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.731466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.731832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.731864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.732209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.732237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.732488] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.732517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.732901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.732931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.733308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.733336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.733568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.733596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.733965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.733995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.734348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.734377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.734727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.734765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.735180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.735208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.735572] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.735601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.735943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.735974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.736328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.736357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.736712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.736740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.737107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.737136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.737504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.737532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.737885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.855 [2024-11-06 09:08:58.737915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.855 qpair failed and we were unable to recover it. 00:33:08.855 [2024-11-06 09:08:58.738290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.738318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.738663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.738691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.739040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.739069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.739312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.739341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.739639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.739667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.739973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.740002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.740307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.740335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.740695] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.740723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.741088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.741117] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.741464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.741493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.741844] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.741873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.742232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.742261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.742562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.742590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.742979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.743008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.743364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.743392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.743745] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.743786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.744148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.744176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.744531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.744566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.744905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.744935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.745294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.745323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.745679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.745708] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.746007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.746037] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.746399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.746427] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.746781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.746811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.747190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.747219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.747580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.747609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.747980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.748010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.748365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.748394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.748744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.748785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh: line 36: 2080145 Killed "${NVMF_APP[@]}" "$@" 00:33:08.856 [2024-11-06 09:08:58.749126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.749155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.749520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.749549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@48 -- # disconnect_init 10.0.0.2 00:33:08.856 [2024-11-06 09:08:58.749914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.749945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:33:08.856 [2024-11-06 09:08:58.750329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.750358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:33:08.856 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:08.856 [2024-11-06 09:08:58.750721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.750760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:08.856 [2024-11-06 09:08:58.750989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.751018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.856 [2024-11-06 09:08:58.751282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.856 [2024-11-06 09:08:58.751311] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.856 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.751673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.751702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.752062] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.752092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.752447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.752477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.752856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.752887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.753226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.753256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.753621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.753657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.753991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.754021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.754437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.754467] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.754810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.754839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.755167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.755198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.755545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.755576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.755916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.755947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.756315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.756346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.756714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.756755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.757101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.757131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.757468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.757499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.757869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.757900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.758254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.758285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.758639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.758670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.758995] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.759027] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@507 -- # nvmfpid=2081059 00:33:08.857 [2024-11-06 09:08:58.759375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@508 -- # waitforlisten 2081059 00:33:08.857 [2024-11-06 09:08:58.759408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:33:08.857 [2024-11-06 09:08:58.759778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@831 -- # '[' -z 2081059 ']' 00:33:08.857 [2024-11-06 09:08:58.759816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.760052] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:08.857 [2024-11-06 09:08:58.760086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:08.857 [2024-11-06 09:08:58.760448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.760480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:08.857 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:08.857 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:08.857 [2024-11-06 09:08:58.760830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.760863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 09:08:58 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:08.857 [2024-11-06 09:08:58.761216] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.761249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.761604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.761636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.761982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.762017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.762399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.762431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.762778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.762811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.763172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.763203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.763562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.763592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.763944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.763977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.857 [2024-11-06 09:08:58.764331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.857 [2024-11-06 09:08:58.764363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.857 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.764604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.764635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.765004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.765035] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.765386] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.765416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.765786] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.765817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.766024] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.766056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.766415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.766446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.766696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.766730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.767029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.767064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.767422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.767454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.767616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.767650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.768053] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.768085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.768443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.768474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.768710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.768740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.769017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.769048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.769400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.769430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.769834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.769866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.770236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.770267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.770502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.770537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.770873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.770905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.771294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.771325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.771716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.771756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.772141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.772171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.772510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.772540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.772922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.772954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.773305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.773337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.773699] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.773730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.774121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.774154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.774519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.774551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.774991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.775023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.775253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.775287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.775629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.775662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.776017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.776052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.776414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.858 [2024-11-06 09:08:58.776446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.858 qpair failed and we were unable to recover it. 00:33:08.858 [2024-11-06 09:08:58.776811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.776850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.777218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.777248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.777611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.777642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.778058] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.778092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.778447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.778478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.778832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.778865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.779131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.779165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.779528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.779558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.779792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.779827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.780203] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.780233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.780609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.780640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.780888] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.780923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.781296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.781327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.781671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.781702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.782079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.782112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.782469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.782500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.782862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.782894] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.783140] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.783174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.783522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.783552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.783793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.783826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.784109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.784140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.784411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.784441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.784811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.784843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.785213] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.785244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.785483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.785513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.785882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.785916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.786313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.786343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.786575] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.786610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.786952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.786984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.787237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.787269] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.787498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.787532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.787808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.787839] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.788215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.788246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.788626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.788657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.789024] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.789056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.789419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.789451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.789797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.789831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.790308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.790339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.859 qpair failed and we were unable to recover it. 00:33:08.859 [2024-11-06 09:08:58.790680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.859 [2024-11-06 09:08:58.790711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.791116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.791150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.791328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.791367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.791661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.791692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.792077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.792111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.792479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.792512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.792881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.792912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.793312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.793343] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.793721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.793765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.794190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.794220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.794599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.794630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.795042] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.795076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.795443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.795473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.795824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.795855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.796233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.796268] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.796616] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.796647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.797073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.797107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.797439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.797470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.797670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.797704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.798118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.798150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.798516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.798546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.798901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.798932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.799279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.799310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.799769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.799802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.800164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.800194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.800550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.800580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.800931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.800963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.801207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.801240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.801630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.801660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.802077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.802110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.802468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.802499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.802825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.802858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.803256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.803287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.803504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.803538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.803907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.803939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.804308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.804339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.804718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.804759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.805127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.805158] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.860 [2024-11-06 09:08:58.805499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.860 [2024-11-06 09:08:58.805530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.860 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.805897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.805929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.806302] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.806332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.806774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.806806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.807189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.807226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.807574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.807606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.807867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.807899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.808273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.808304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.808672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.808706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.809082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.809114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.809509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.809542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.809904] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.809936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.810321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.810354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.810712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.810744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.811026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.811056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.811411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.811442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.811815] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.811847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.812202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.812234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.812608] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.812640] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.812981] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.813014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.813399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.813431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.813778] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.813810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.814046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.814078] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.814528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.814560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.814923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.814954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.815233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.815267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.815639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.815671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.816040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.816074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.816416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.816447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.816809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.816842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.817233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.817266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.817653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.817687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.817924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.817956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.818312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.818344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.818763] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.818795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.819149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.819180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.819564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.819594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.819833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.819865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.820282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.861 [2024-11-06 09:08:58.820312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.861 qpair failed and we were unable to recover it. 00:33:08.861 [2024-11-06 09:08:58.820682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.820714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.821040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.821073] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.821436] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.821468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.821812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.821845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.822181] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.822211] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 [2024-11-06 09:08:58.822214] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.822276] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:08.862 [2024-11-06 09:08:58.822604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.822636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.823029] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.823059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.823405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.823436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.823803] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.823835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.824205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.824236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.824593] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.824624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.824977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.825010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.825264] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.825296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.825665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.825698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.826093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.826126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.826528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.826560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.826935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.826967] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.827363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.827407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.827770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.827804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.828164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.828195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.828555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.828586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.828935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.828968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.829328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.829360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.829743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.829788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.830165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.830202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.830570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.830602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.830849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.830884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.831165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.831197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.831461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.831496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.831885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.831919] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.832185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.832215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.832468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.832502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.832856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.832890] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.833260] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.833292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.833427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.833458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.833827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.833860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.834249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.834280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.862 qpair failed and we were unable to recover it. 00:33:08.862 [2024-11-06 09:08:58.834654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.862 [2024-11-06 09:08:58.834688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.835062] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.835095] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.835431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.835463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.835705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.835736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.835974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.836007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.836252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.836288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.836653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.836684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.837111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.837147] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.837485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.837519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.837785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.837820] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.838231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.838263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.838611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.838642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.838987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.839020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.839379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.839411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.839797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.839830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.840107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.840139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.840544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.840576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.840928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.840962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.841336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.841367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.841731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.841789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.842157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.842195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.842430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.842460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.842839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.842871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.843240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.843273] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.843668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.843700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.843948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.843981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.844263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.844295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.844532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.844563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.844908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.844939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.845174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.845207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.845461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.845496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.845873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.845905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.846304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.846335] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.846690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.863 [2024-11-06 09:08:58.846722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.863 qpair failed and we were unable to recover it. 00:33:08.863 [2024-11-06 09:08:58.847036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.847069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.847458] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.847489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.847855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.847889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.848252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.848284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.848682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.848714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.849093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.849126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.849465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.849496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.849728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.849772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.850145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.850177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.850423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.850458] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.850678] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.850710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.851059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.851092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.851466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.851498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.851859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.851891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.852124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.852157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.852530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.852562] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.852936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.852970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.853340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.853371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.853739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.853782] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.854137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.854168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.854570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.854601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.854847] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.854879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.855270] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.855302] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.855662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.855694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.856037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.856070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.856430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.856462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.856833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.856872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.857126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.857157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.857546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.857577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.857926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.857958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.858331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.858362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.858758] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.858791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.859185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.859217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.859599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.859630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.859884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.859917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.860289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.860321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.860671] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.860702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.861111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.864 [2024-11-06 09:08:58.861143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.864 qpair failed and we were unable to recover it. 00:33:08.864 [2024-11-06 09:08:58.861495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.861527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.861886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.861918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.862300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.862332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.862735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.862780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.863045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.863075] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.863478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.863509] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.863796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.863830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.864218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.864250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.864617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.864648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.864993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.865026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.865383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.865414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.865720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.865760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.866118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.866150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.866515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.866546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.866911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.866945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.867331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.867363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.867631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.867662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.868124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.868156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.868425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.868456] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.868833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.868866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.869256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.869287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.869669] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.869700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.870075] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.870108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.870481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.870511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.870895] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.870927] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.871301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.871332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.871731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.871787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.872057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.872089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.872459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.872498] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.872873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.872907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.873261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.873291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.873602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.873633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.874043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.874074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.874433] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.874464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.874837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.874869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.875234] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.875266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.875633] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.875665] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.876105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.876137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.865 [2024-11-06 09:08:58.876405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.865 [2024-11-06 09:08:58.876436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.865 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.876835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.876867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.877261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.877293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.877656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.877688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.877955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.877988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.878416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.878448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.878789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.878822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.879199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.879231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.879596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.879628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.879972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.880004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.880386] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.880419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.880792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.880825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.881221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.881253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.881651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.881681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.881971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.882003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.882355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.882386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.882621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.882653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.883026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.883059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.883457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.883490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.883733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.883777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.884147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.884179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.884534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.884566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.884937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.884970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.885332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.885363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.885717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.885765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.886204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.886237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.886476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.886508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.886770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.886801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.887051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.887087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.887439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.887472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.887742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.887790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.888178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.888210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.888596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.888628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.888871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.888906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.889311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.889342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.889568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.889602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.889862] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.889897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.890254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.890286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.866 [2024-11-06 09:08:58.890666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.866 [2024-11-06 09:08:58.890698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.866 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.891093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.891125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.891492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.891524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.891891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.891925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.892299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.892331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.892682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.892715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.893092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.893125] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.893484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.893516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.893896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.893928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.894141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.894176] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.894587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.894619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.894993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.895026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.895390] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.895422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.895794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.895825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.896184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.896215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.896562] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.896595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.896967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.897001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.897137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.897173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.897523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.897556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.897905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.897945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.898272] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.898304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.898646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.898678] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.898942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.898979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.899340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.899372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.899726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.899770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.900143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.900174] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.900555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.900586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.900933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.900966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.901376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.901407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.901791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.901825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.902187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.902218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.902595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.902626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.902982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.903014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.903360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.867 [2024-11-06 09:08:58.903391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.867 qpair failed and we were unable to recover it. 00:33:08.867 [2024-11-06 09:08:58.903777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.903811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.904049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.904082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.904449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.904479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.904839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.904872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.905257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.905288] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.905657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.905688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.906066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.906099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.906456] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.906487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.906867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.906899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.907275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.907306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.907711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.907742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.908039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.908071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.908326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.908361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.908719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.908762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.909149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.909179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.909553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.909586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.909959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.909994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.910363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.910395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.910770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.910803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.911176] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.911207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.911590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.911621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.911993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.912026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.912371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.912402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.912740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.912787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.913198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.913228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.913591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.913629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.913965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.913997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.914351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.914383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.914715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.914760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.915153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.915185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.915556] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.915589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.915922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.915954] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.916173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.916205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.916561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.916592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.916985] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.917016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.917395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.917428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.917681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.917715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.918095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.918126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.868 qpair failed and we were unable to recover it. 00:33:08.868 [2024-11-06 09:08:58.918484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.868 [2024-11-06 09:08:58.918516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.918927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.918960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.919341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.919373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.919779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.919813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.920195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.920227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.920596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.920634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.920986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.921018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.921389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.921420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.921803] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.921838] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.922208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.922239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.922628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.922661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.923022] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.923058] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.923419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.923450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.923814] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.923846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.924227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.924258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.924642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.924673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.925066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.925098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.925449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.925480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.925713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.925744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.926139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.926170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.926536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.926567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.926930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.926963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.927346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.927377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.927737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.927779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.928101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.928131] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.928503] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.928534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.928908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.928940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.929000] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:33:08.869 [2024-11-06 09:08:58.929318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.929349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.929727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.929780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.930174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.930205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.930454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.930484] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.930856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.930889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.931150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.931184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.931484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.931516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.931907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.931939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.932332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.932363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.932789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.932821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.933172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.933203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.869 qpair failed and we were unable to recover it. 00:33:08.869 [2024-11-06 09:08:58.933604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.869 [2024-11-06 09:08:58.933636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.934024] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.934059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.934422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.934460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.934772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.934805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.935050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.935080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.935217] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.935246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.935658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.935689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.936158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.936191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.936573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.936605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.936961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.936995] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.937387] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.937418] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.937781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.937815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.938229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.938261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.938490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.938521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.938900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.938933] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.939303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.939334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.939720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.939762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.940154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.940186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.940539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.940572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.940936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.940969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.941349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.941380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.941741] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.941801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.942166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.942197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.942546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.942577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.942967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.943002] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.943301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.943333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.943700] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.943730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.944119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.944152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.944499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.944529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.944908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.944941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.945329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.945361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.945723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.945777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.946153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.946184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.946540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.946570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.946827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.946860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.947220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.947252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.947622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.947654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.948081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.948113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:08.870 [2024-11-06 09:08:58.948352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:08.870 [2024-11-06 09:08:58.948387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:08.870 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.948769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.948804] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.949145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.949177] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.949521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.949553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.949799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.949842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.950227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.950259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.950624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.950656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.952782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.952856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.953356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.953394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.953772] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.953805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.954172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.954205] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.954568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.954599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.954945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.954982] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.149 [2024-11-06 09:08:58.955343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.149 [2024-11-06 09:08:58.955375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.149 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.955780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.955813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.956188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.956219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.956597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.956630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.956993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.957025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.957437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.957468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.957827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.957861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.958209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.958240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.958592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.958623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.958998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.959030] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.959394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.959426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.959668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.959703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.960100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.960133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.960468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.960500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.960851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.960884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.961267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.961297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.961636] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.961667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.962120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.962152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.962395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.962431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.962806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.962840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.963204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.963235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.963620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.963651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.964034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.964069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.964349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.964379] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.964615] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.964651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.965033] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.965066] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.965423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.965454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.965723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.965767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.966145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.966175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.966416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.966450] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.966809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.966842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.967236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.967275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.967623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.967653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.967900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.967931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.968307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.968339] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.968760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.968791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.969152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.969182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.969510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.969542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.969972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.970006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.970322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.970352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.970709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.970739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.971113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.971146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.971512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.971542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.971906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.971939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.972294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.972326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.972703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.972734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.973136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.973167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.973535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.973567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.973921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.973953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.974199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.974232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.974597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.974629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.974977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.975009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.975362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.975392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.975768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.975801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.976167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.976198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.976564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.976594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.976931] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.976964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.977301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.977333] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.977738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.977785] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.978156] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.978187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.978559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.978592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.978940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.978972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.150 qpair failed and we were unable to recover it. 00:33:09.150 [2024-11-06 09:08:58.979210] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.150 [2024-11-06 09:08:58.979244] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.979635] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.979667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.980036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.980069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.980437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.980468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.980833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.980867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.981244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.981275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.981440] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:33:09.151 [2024-11-06 09:08:58.981485] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:33:09.151 [2024-11-06 09:08:58.981493] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:33:09.151 [2024-11-06 09:08:58.981501] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:33:09.151 [2024-11-06 09:08:58.981507] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:33:09.151 [2024-11-06 09:08:58.981625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.981656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.982006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.982045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.982381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.982413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.982783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.982817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.983209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.983241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.983462] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:33:09.151 [2024-11-06 09:08:58.983619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.983649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.983615] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:33:09.151 [2024-11-06 09:08:58.983805] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:33:09.151 [2024-11-06 09:08:58.983806] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:33:09.151 [2024-11-06 09:08:58.983998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.984029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.984288] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.984318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.984670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.984702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.985087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.985120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.985492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.985523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.985886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.985917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.986291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.986323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.986696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.986730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.987123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.987155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.987413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.987444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.987789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.987823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.988194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.988227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.988522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.988554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.988852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.988884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.989252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.989283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.989651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.989681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.990051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.990084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.990464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.990497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.990843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.990874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.991259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.991290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.991640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.991673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.991973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.992005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.992253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.992287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.992539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.992571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.992940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.992973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.993305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.993336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.993689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.993720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.994076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.994108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.994464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.994496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.994870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.994902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.995281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.995313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.995692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.995723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.996119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.996152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.996425] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.996455] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.996827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.996868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.997246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.997278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.997627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.997660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.997999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.998033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.998375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.998407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.998739] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.998786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.999148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.999180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.999557] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.999590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:58.999946] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:58.999979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:59.000329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:59.000361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:59.000719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:59.000777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:59.001111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:59.001143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.151 qpair failed and we were unable to recover it. 00:33:09.151 [2024-11-06 09:08:59.001497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.151 [2024-11-06 09:08:59.001529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.001872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.001906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.002293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.002327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.002718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.002761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.003117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.003149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.003531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.003563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.003926] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.003959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.004359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.004392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.004774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.004810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.005185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.005218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.005476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.005508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.005837] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.005870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.006244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.006277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.006659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.006691] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.007082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.007115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.007494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.007526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.007886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.007920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.008244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.008276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.008629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.008662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.008992] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.009025] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.009376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.009409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.009681] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.009714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.010081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.010116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.010480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.010513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.010911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.010947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.011354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.011385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.011735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.011789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.012163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.012195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.012433] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.012475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.012846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.012881] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.013259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.013293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.013663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.013696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.014066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.014100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.014343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.014378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.014694] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.014727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.014987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.015020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.015367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.015400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.015769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.015803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.016155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.016187] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.016510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.016542] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.016791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.016823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.017198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.017230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.017588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.017620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.017967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.017999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.018375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.018407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.018779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.018813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.019166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.019198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.019573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.019606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.020061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.020096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.020473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.020505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.020866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.020899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.021274] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.021306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.021647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.021679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.022037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.022072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.152 qpair failed and we were unable to recover it. 00:33:09.152 [2024-11-06 09:08:59.022430] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.152 [2024-11-06 09:08:59.022465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.022703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.022734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.023103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.023136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.023464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.023497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.023853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.023886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.024268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.024300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.024654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.024686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.025076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.025110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.025461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.025494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.025720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.025765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.026044] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.026074] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.026461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.026494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.026846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.026885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.027265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.027297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.027520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.027567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.027901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.027935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.028330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.028361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.028589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.028623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.028842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.028875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.029249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.029280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.029632] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.029664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.030007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.030040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.030252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.030284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.030517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.030549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.030818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.030854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.031252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.031283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.031640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.031672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.031916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.031948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.032290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.032321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.032573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.032605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.032730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.032779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.033132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.033165] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.033517] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.033549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.033878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.033912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.034136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.034168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.034508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.034540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.034898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.034930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.035294] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.035325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.035686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.035718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.036073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.036105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.036450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.036481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.036728] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.036783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.037197] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.037229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.037441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.037475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.037827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.037861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.038250] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.038282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.038568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.038599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.038833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.038866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.039221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.039253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.039627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.039660] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.040054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.040087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.040417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.040449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.040672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.040704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.041090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.041122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.041267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.041309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.041531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.041563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.041846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.041877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.042266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.042298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.042645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.042676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.042912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.042945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.043170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.153 [2024-11-06 09:08:59.043206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.153 qpair failed and we were unable to recover it. 00:33:09.153 [2024-11-06 09:08:59.043538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.043569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.043809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.043841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.044076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.044107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.044367] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.044402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.044650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.044681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.044901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.044935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.045333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.045365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.045733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.045781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.046114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.046146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.046501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.046532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.046773] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.046806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.047202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.047234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.047601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.047633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.048007] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.048038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.048407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.048438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.048540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.048570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.048914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.048945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.049227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.049258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.049630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.049663] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.049894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.049929] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.050168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.050201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.050440] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.050471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.050587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.050619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.050856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.050888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.051221] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.051253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.051592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.051623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.051855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.051888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.052135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.052166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.052386] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.052416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.052661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.052696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.052977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.053009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.053359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.053390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.053780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.053813] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.054163] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.054201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.054574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.054606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.054984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.055017] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.055244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.055274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.055528] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.055559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.055785] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.055816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.056162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.056193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.056527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.056558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.056905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.056937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.057179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.057210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.057468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.057500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.057759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.057794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.057898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.057925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6094000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.058270] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x845e00 is same with the state(6) to be set 00:33:09.154 [2024-11-06 09:08:59.058867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.058998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.059445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.059486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.059835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.059873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.059991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.060020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.060154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.060183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.060393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.060426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.060810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.060843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.061248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.061279] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.061495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.061526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.061905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.061938] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.062333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.062363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.062732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.062774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.063190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.063221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.063545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.063586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.063942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.063975] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.064342] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.064373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.064760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.154 [2024-11-06 09:08:59.064793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.154 qpair failed and we were unable to recover it. 00:33:09.154 [2024-11-06 09:08:59.065054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.065091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.065421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.065453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.065784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.065816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.066212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.066243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.066614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.066647] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.067001] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.067033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.067462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.067493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.067863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.067895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.068276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.068306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.068674] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.068706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.069138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.069171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.069539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.069571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.069827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.069859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.070186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.070216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.070583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.070613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.070973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.071004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.071331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.071362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.071730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.071769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.072057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.072088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.072338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.072369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.072760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.072794] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.073187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.073217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.073568] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.073600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.073978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.074014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.074384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.074416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.074673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.074705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.074964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.074997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.075385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.075415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.075812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.075846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.076250] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.076283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.076655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.076686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.077047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.077079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.077356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.077387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.077713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.077744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.077991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.078021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.078285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.078318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.078690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.078723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.079151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.079182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.079543] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.079574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.079933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.079965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.080299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.080330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.080670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.080701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.080961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.080996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.081379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.081409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.081783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.081816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.082196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.082229] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.082555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.082587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.082939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.082971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.083353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.083383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.083705] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.083737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.083904] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.083937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.084207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.084239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.084606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.084638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.084977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.085010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.085364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.155 [2024-11-06 09:08:59.085394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.155 qpair failed and we were unable to recover it. 00:33:09.155 [2024-11-06 09:08:59.085722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.085778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.086115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.086146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.086512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.086543] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.086933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.086966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.087297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.087328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.087651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.087682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.088039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.088072] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.088446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.088478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.088831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.088870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.089244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.089275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.089635] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.089667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.090061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.090093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.090463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.090494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.090838] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.090871] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.091257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.091287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.091680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.091713] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.092087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.092119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.092469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.092500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.092724] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.092767] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.093131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.093162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.093518] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.093550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.093886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.093918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.094242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.094274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.094631] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.094662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.095006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.095039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.095419] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.095449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.095827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.095859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.096097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.096128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.096497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.096529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.096875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.096907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.097252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.097283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.097611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.097642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.097999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.098031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.098437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.098468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.098828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.098861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.099258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.099290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.099622] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.099653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.100013] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.100046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.100408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.100439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.100782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.100815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.101205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.101238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.101601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.101632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.101982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.102014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.102347] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.102378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.102775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.102809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.103209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.103240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.103607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.103638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.103961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.103993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.104363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.104400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.104720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.104759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.105144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.105175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.105569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.105602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.105970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.106004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.106352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.106383] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.106767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.106801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.107026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.107057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.107311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.107341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.107600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.107633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.107843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.107877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.108232] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.108263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.108510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.108541] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.156 [2024-11-06 09:08:59.108878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.156 [2024-11-06 09:08:59.108910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.156 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.109148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.109182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.109512] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.109544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.109806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.109842] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.110108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.110140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.110502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.110533] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.110781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.110812] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.111188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.111218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.111460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.111494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.111819] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.111852] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.112105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.112140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.112384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.112415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.112637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.112668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.113033] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.113064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.113451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.113483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.113843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.113875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.114190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.114221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.114322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.114351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.114587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.114619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.114962] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.114994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.115353] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.115384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.115768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.115800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.116028] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.116059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.116465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.116496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.116867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.116900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.117275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.117306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.117555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.117589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.117725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.117776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.118036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.118069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.118421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.118452] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.118767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.118801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.119153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.119185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.119496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.119526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.119869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.119902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.120132] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.120161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.120380] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.120410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.120768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.120802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.121167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.121203] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.121414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.121445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.121689] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.121724] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.121967] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.122000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.122379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.122410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.122627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.122657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.123025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.123059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.123426] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.123457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.123834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.123866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.124099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.124134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.124346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.124377] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.124779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.124810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.125191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.125222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.125565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.125596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.125956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.125989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.126227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.126258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.126627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.126658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.127011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.127044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.127431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.127463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.127686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.127718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.128083] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.128115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.128481] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.128511] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.128871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.128904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.129292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.129323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.157 [2024-11-06 09:08:59.129692] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.157 [2024-11-06 09:08:59.129723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.157 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.129965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.129998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.130374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.130404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.130790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.130824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.131191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.131222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.131592] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.131623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.131719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.131768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.132110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.132141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.132510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.132544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.132765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.132799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.133026] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.133056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.133279] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.133310] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.133525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.133556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.133903] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.133936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.134080] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.134109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.134489] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.134521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.134789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.134824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.135185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.135218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.135600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.135631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.135966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.135997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.136218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.136250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.136473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.136504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.136710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.136741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.137104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.137136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.137305] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.137337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.137702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.137734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.138090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.138121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.138472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.138503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.138600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.138631] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.138969] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.139001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.139350] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.139382] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.139762] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.139793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.140142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.140173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.140530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.140560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.140933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.140965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.141366] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.141398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.141607] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.141637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.141999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.142031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.142240] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.142271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.142490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.142520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.142771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.142803] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.143188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.143219] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.143434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.143465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.143877] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.143910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.144150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.144182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.144473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.144504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.144720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.144765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.145112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.145143] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.145514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.145545] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.145929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.145960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.146332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.146364] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.146708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.146739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.147104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.147135] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.147503] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.147535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.147779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.147815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.158 [2024-11-06 09:08:59.147921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.158 [2024-11-06 09:08:59.147950] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.158 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.148292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.148322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.148679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.148711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.149067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.149099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.149314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.149345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.149691] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.149722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.150056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.150088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.150320] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.150351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.150707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.150738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.150988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.151019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.151247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.151277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.151507] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.151539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.151915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.151948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.152182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.152217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.152571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.152602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.152974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.153007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.153360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.153390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.153727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.153763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.154006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.154038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.154423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.154454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.154830] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.154862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.155220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.155251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.155583] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.155615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.155832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.155863] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.156229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.156259] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.156584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.156617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.156835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.156867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.157245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.157275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.157653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.157684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.157905] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.157936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.158187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.158218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.158558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.158595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.158933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.158965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.159309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.159341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.159550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.159581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.159935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.159969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.160360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.160391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.160623] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.160657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.161002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.161033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.161363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.161395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.161718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.161757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.162102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.162132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.162359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.162391] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.162626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.162657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.162896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.162930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.163236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.163267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.163640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.163672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.163916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.163948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.164340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.164371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.164590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.164622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.164978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.165009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.165399] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.165431] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.165665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.165698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.165989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.166022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.166406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.166437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.166767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.166800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.167158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.167188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.167551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.167581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.167801] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.167836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.168167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.168199] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.168577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.168609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.168708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.168740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.169111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.169142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.169292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.159 [2024-11-06 09:08:59.169325] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.159 qpair failed and we were unable to recover it. 00:33:09.159 [2024-11-06 09:08:59.169704] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.169734] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.170093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.170124] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.170496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.170528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.170908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.170940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.171151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.171183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.171582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.171613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.171965] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.171998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.172207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.172246] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.172490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.172521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.172868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.172900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.173254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.173285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.173653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.173684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.174039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.174070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.174423] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.174453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.174822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.174854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.175119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.175149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.175369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.175400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.175769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.175801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.176128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.176157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.176365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.176395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.176722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.176762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.177021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.177053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.177417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.177448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.177775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.177807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.178222] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.178253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.178603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.178633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.178972] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.179004] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.179377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.179407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.179709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.179741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.179988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.180019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.180402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.180433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.180783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.180815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.181211] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.181241] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.181614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.181646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.182020] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.182054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.182444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.182474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.182687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.182717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.183077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.183109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.183448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.183479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.183703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.183735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.184121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.184153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.184536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.184567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.184784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.184817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.185173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.185204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.185421] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.185451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.185829] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.185862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.186003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.186033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.186401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.186437] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.186646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.186679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.187030] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.187063] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.187393] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.187425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.187743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.187784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.188139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.188169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.188495] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.188525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.188927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.188959] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.189307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.189337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.189673] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.189703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.190066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.190098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.190445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.190475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.190850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.190883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.191283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.160 [2024-11-06 09:08:59.191313] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.160 qpair failed and we were unable to recover it. 00:33:09.160 [2024-11-06 09:08:59.191670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.191702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.192078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.192110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.192493] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.192524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.192884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.192917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.193299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.193330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.193564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.193595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.193852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.193886] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.194244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.194275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.194647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.194679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.195014] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.195046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.195372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.195404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.195781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.195814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.196166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.196196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.196577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.196610] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.197023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.197055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.197422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.197453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.197817] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.197849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.198184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.198215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.198472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.198505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.198824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.198857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.199185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.199216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.199570] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.199602] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.199971] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.200003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.200401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.200434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.200784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.200816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.201057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.201087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.201345] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.201384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.201727] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.201769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.202088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.202119] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.202491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.202524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.202882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.202916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.203295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.203326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.203707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.203738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.203999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.204031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.204377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.204409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.204666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.204697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.205046] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.205079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.205472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.205502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.205879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.205911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.206286] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.206318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.206690] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.206723] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.207110] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.207142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.207514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.207547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.207892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.207925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.208259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.208290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.208638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.208669] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.209010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.209042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.209412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.209443] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.209653] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.209684] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.209935] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.209968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.210324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.210355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.210723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.210763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.211098] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.211130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.211501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.211532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.161 [2024-11-06 09:08:59.211882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.161 [2024-11-06 09:08:59.211918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.161 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.212275] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.212307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.212630] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.212662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.213027] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.213060] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.213388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.213420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.213784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.213818] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.214211] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.214242] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.214580] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.214614] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.214966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.214999] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.215352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.215384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.215779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.215815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.216147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.216180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.216532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.216570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.216832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.216867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.217225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.217257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.217605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.217636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.218006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.218039] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.218370] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.218401] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.218765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.218829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.219223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.219256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.219618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.219650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.220010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.220043] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.220410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.220441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.220812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.220845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.221166] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.221197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.221451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.221482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.221863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.221895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.222277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.222308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.222647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.222679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.223049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.223082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.223439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.223470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.223839] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.223872] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.224225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.224257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.224617] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.224648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.225003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.225036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.225428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.225459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.225813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.225845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.226184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.226215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.226603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.226634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.226759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.226793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.227039] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.227071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.227402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.227434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.227803] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.227836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.228226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.228257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.228478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.228510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.228716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.228756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.229078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.229108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.229326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.229357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.229708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.229740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.230111] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.230142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.230355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.230386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.230767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.230799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.231193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.231231] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.231596] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.231627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.232000] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.232033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.232395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.232426] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.232794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.232826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.233173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.233204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.233466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.233500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.233714] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.233756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.234045] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.162 [2024-11-06 09:08:59.234076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.162 qpair failed and we were unable to recover it. 00:33:09.162 [2024-11-06 09:08:59.234281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.234312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.234547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.234578] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.234900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.234931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.235325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.235355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.235569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.235600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.235978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.236010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.236341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.236372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.236587] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.236616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.236841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.236873] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.237273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.237304] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.237520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.237551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.237779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.237810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.238219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.238250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.238620] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.238652] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.238993] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.239031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.239408] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.239440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.239787] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.239819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.240057] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.240092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.240467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.240500] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.240715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.240757] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.241015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.241046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.241451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.241483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.241868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.241899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.242254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.242284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.242656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.242687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.242835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.242869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.243096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.243127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.243490] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.243521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.243892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.243924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.244129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.244160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.163 [2024-11-06 09:08:59.244519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.163 [2024-11-06 09:08:59.244549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.163 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.244934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.244977] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.245204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.245236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.245639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.245671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.246036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.246068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.246429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.246461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.246684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.246714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.247082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.247114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.247299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.247331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.247576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.247607] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.247852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.247889] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.248239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.248271] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.248643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.248674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.248963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.248994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.249335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.249366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.249738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.249780] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.249897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.249930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.250193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.250226] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.250590] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.250621] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.250840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.250874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.251251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.251282] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.251499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.251529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.251866] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.251899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.252141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.252178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.252527] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.252559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.439 qpair failed and we were unable to recover it. 00:33:09.439 [2024-11-06 09:08:59.252934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.439 [2024-11-06 09:08:59.252965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.253177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.253208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.253565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.253597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.253812] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.253844] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.254207] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.254238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.254457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.254489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.254861] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.254893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.255244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.255274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.255641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.255672] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.255898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.255930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.256276] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.256308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.256523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.256554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.256770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.256802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.257023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.257054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.257385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.257416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.257663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.257693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.257943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.257976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.258204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.258235] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.258477] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.258507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.258867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.258900] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.259267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.259298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.259679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.259709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.259986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.260021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.260384] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.260415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.260537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.260571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.260873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.260905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.261133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.261164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.261532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.261565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.261968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.262000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.262218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.262248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.262479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.262510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.262868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.262901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.263263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.263293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.263520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.263555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.263891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.263923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.264167] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.264198] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.264442] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.264475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.264846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.264877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.265115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.265146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.265392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.265425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.265777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.265810] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.266151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.266184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.266559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.266589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.266940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.266979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.267358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.267390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.267766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.267799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.268190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.268222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.268589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.268620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.268988] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.269020] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.269366] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.269398] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.269611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.269641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.270010] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.270042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.270402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.270434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.270802] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.270835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.271177] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.271208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.271559] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.271590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.271937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.271968] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.440 qpair failed and we were unable to recover it. 00:33:09.440 [2024-11-06 09:08:59.272376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.440 [2024-11-06 09:08:59.272408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.272765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.272798] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.273101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.273132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.273503] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.273534] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.273759] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.273791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.274158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.274188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.274563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.274593] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.274979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.275012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.275344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.275376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.275760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.275791] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.276153] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.276183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.276565] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.276597] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.276937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.276969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.277341] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.277373] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.277755] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.277789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.278178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.278207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.278560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.278591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.278925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.278958] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.279281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.279312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.279670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.279702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.280055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.280088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.280403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.280433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.280787] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.280819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.281201] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.281234] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.281561] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.281591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.281846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.281879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.282226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.282262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.282636] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.282668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.283034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.283067] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.283470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.283501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.283873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.283905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.284268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.284298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.284667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.284699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.285055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.285088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.285443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.285475] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.285799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.285830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.286220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.286252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.286611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.286642] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.286976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.287008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.287415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.287445] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.287803] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.287835] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.288056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.288086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.288463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.288493] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.288842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.288875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.289261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.289292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.289665] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.289696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.290090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.290122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.290432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.290463] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.290769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.290801] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.291160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.291192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.291576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.291608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.291947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.291979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.292351] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.292381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.292757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.292790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.293144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.293175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.293524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.293555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.293925] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.293956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.441 [2024-11-06 09:08:59.294326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.441 [2024-11-06 09:08:59.294356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.441 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.294691] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.294721] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.295084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.295116] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.295483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.295515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.295889] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.295921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.296299] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.296331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.296698] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.296729] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.297082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.297113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.297441] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.297472] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.297857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.297895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.298148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.298179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.298552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.298583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.298937] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.298969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.299325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.299355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.299564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.299595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.299983] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.300014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.300363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.300393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.300718] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.300759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.301127] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.301157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.301508] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.301539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.301915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.301947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.302311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.302341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.302708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.302740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.303133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.303166] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.303540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.303571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.303941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.303973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.304324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.304356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.304670] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.304701] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.304811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.304841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.305202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.305233] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.305470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.305505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.305915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.305947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.306162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.306193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.306405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.306436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.306791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.306822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.307073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.307105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.307479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.307512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.307872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.307905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.308277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.308308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.308687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.308718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.308999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.309034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.309244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.309277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.309503] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.309535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.309941] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.309974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.310323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.310354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.310726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.310768] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.311020] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.311053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.311285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.311316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.311647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.311679] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.312048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.312093] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.312463] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.442 [2024-11-06 09:08:59.312494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.442 qpair failed and we were unable to recover it. 00:33:09.442 [2024-11-06 09:08:59.312595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.312623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.313016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.313047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.313404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.313434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.313809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.313841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.314068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.314099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.314499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.314530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.314753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.314784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.315008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.315038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.315445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.315476] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.315850] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.315882] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.316119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.316150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.316498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.316530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.316764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.316796] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.317017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.317047] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.317451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.317482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.317855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.317887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.318271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.318301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.318626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.318658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.319004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.319036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.319403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.319434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.319799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.319836] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.320072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.320105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.320478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.320510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.320734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.320774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.321012] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.321042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.321415] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.321449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.321777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.321808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.322193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.322224] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.322594] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.322626] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.322921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.322953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.323356] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.323387] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.323736] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.323781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.324175] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.324206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.324418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.324448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.324799] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.324831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.325215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.325247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.325457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.325488] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.325841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.325874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.326199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.326236] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.326603] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.326635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.327017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.327048] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.327422] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.327453] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.327576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.327616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.327921] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.327952] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.328178] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.328209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.328497] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.328528] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.328881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.328915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.329126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.329157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.329386] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.329417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.329783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.329814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.330157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.330189] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.330558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.330591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.330928] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.330960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.331189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.331220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.331638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.331670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.332049] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.332081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.332404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.332434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.443 qpair failed and we were unable to recover it. 00:33:09.443 [2024-11-06 09:08:59.332657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.443 [2024-11-06 09:08:59.332690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.333047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.333079] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.333464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.333495] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.333768] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.333802] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.334165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.334197] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.334530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.334560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.334915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.334947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.335324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.335356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.335460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.335491] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.335852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.335885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.336246] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.336278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.336493] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.336523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.336920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.336951] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.337318] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.337349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.337563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.337594] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.337932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.337964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.338192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.338225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.338510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.338540] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.338923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.338956] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.339176] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.339208] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.339576] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.339608] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.339984] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.340022] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.340371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.340403] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.340789] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.340821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.341031] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.341062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.341433] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.341462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.341793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.341826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.342219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.342250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.342499] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.342530] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.342867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.342899] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.343003] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.343033] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.343389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.343420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.343803] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.343834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.344185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.344216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.344462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.344494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.344825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.344858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.345209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.345240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.345573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.345605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.345968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.346000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.346407] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.346438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.346795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.346827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.347194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.347225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.347602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.347635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.348011] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.348042] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.348371] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.348402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.348777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.348809] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.349187] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.349218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.349598] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.349630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.349973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.350006] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.350363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.350394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.350725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.350765] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.351121] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.351153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.351545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.351577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.351800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.351832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.352179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.352210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.352560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.352592] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.352923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.352955] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.353324] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.444 [2024-11-06 09:08:59.353355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.444 qpair failed and we were unable to recover it. 00:33:09.444 [2024-11-06 09:08:59.353707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.353739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.354120] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.354152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.354487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.354519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.354899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.354937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.355169] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.355200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.355588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.355619] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.355974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.356008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.356218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.356249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.356584] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.356615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.356994] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.357026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.357400] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.357430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.357765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.357797] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.358164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.358195] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.358534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.358565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.358947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.358978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.359337] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.359368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.359744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.359814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.360202] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.360232] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.360585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.360617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.360969] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.361003] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.361383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.361413] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.361779] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.361811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.362184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.362215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.362586] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.362617] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.362948] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.362979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.363304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.363336] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.363661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.363693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.363953] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.363984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.364377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.364409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.364757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.364789] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.365155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.365186] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.365564] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.365596] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.365934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.365966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.366332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.366362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.366734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.366774] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.367157] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.367188] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.367548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.367580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.367964] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.367998] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.368368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.368399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.368767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.368799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.369161] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.369192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.369523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.369555] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.369916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.369948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.370309] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.370347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.370680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.370711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.371105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.371136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.371514] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.371544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.371893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.371925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.372307] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.372338] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.372711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.372741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.373126] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.373157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.373541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.445 [2024-11-06 09:08:59.373573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.445 qpair failed and we were unable to recover it. 00:33:09.445 [2024-11-06 09:08:59.373949] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.373980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.374365] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.374396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.374738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.374779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.375095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.375127] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.375464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.375494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.375821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.375854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.376233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.376263] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.376645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.376677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.377036] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.377069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.377398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.377430] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.377774] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.377805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.378171] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.378202] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.378573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.378604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.378938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.378969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.379339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.379371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.379698] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.379731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.380104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.380134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.380498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.380529] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.380871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.380904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.381274] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.381306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.381555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.381587] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.381912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.381945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.382334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.382366] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.382743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.382781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.383142] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.383173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.383551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.383582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.383945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.383978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.384377] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.384408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.384742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.384790] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.385074] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.385106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.385429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.385460] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.385834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.385879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.386257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.386289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.386618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.386649] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.386987] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.387018] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.387386] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.387416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.387760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.387793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.388151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.388182] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.388550] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.388581] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.388809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.388843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.389219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.389249] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.389591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.389622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.389990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.390023] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.390360] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.390390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.390764] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.390795] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.391136] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.391169] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.391540] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.391570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.391933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.391965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.392293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.392326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.392539] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.392570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.392939] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.392971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.393363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.393395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.393610] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.393641] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.393979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.394011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.394406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.394438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.394650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.394681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.395037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.395069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.446 [2024-11-06 09:08:59.395455] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.446 [2024-11-06 09:08:59.395486] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.446 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.395857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.395891] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.396259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.396290] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.396606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.396636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.396864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.396896] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.397268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.397299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.397409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.397441] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.397798] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.397831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.398214] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.398247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.398619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.398650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.399025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.399057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.399303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.399334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.399716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.399755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.400114] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.400145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.400525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.400563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.400893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.400924] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.401314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.401345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.401696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.401728] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.402112] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.402145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.402358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.402390] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.402744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.402787] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.403108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.403139] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.403531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.403563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.403870] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.403901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.404082] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.404113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.404322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.404353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.404719] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.404758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.405084] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.405115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.405492] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.405523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.405874] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.405906] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.406295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.406326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.406654] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.406685] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.406947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.406981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.407193] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.407225] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.407431] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.407462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.407824] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.407857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.408236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.408267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.408476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.408507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.408879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.408912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.409268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.409300] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.409552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.409583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.409930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.409963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.410314] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.410344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.410675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.410706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.411081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.411114] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.411446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.411478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.411585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.411615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.411852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.411884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.412283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.412314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.412549] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.412584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.412808] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.412840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.413182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.413214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.413546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.413577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.413792] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.413824] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.414047] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.414084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.414362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.414394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.414745] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.414784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.414998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.415028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.447 qpair failed and we were unable to recover it. 00:33:09.447 [2024-11-06 09:08:59.415290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.447 [2024-11-06 09:08:59.415322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.415535] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.415567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.415800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.415834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.416165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.416196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.416553] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.416584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.416959] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.416993] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.417335] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.417367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.417601] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.417635] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.417852] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.417884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.418233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.418265] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.418633] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.418666] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.418901] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.418936] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.419312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.419342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.419744] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.419786] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.419891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.419921] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.420137] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.420168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.420519] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.420552] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.420884] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.420916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.421319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.421350] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.421577] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.421609] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.421990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.422021] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.422344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.422375] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.422730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.422770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.422991] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.423024] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.423403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.423435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.423784] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.423815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.424155] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.424185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.424418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.424448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.424662] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.424693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.424940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.424972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.425330] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.425361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.425574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.425606] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.425950] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.425981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.426328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.426360] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.426571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.426601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.426982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.427012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.427395] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.427433] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.427797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.427829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.428219] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.428250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.428468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.428499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.428856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.428888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.429254] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.429285] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.429656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.429688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.430072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.430106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.430475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.430505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.430882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.430915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.431281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.431312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.431643] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.431674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.432038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.432070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.432434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.432466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.432794] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.432827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.433073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.433104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.448 qpair failed and we were unable to recover it. 00:33:09.448 [2024-11-06 09:08:59.433462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.448 [2024-11-06 09:08:59.433494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.433822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.433856] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.434190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.434222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.434484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.434520] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.434846] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.434878] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.435236] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.435266] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.435621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.435653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.435982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.436014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.436339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.436372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.436720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.436758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.437081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.437112] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.437524] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.437556] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.437914] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.437946] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.438312] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.438342] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.438726] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.438769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.439004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.439034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.439266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.439297] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.439706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.439736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.440139] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.440171] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.440546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.440577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.440904] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.440937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.441289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.441320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.441708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.441738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.442076] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.442107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.442465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.442502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.442879] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.442912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.443289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.443321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.443682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.443714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.444059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.444091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.444424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.444454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.444831] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.444864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.445249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.445281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.445667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.445700] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.446066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.446100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.446475] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.446508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.446864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.446897] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.447271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.447305] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.447632] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.447664] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.447924] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.447960] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.448329] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.448362] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.448733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.448776] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.449182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.449215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.449464] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.449497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.449872] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.449904] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.450296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.450328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.450682] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.450715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.451050] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.451082] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.451468] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.451499] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.451835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.451868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.452255] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.452286] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.452660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.452692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.452996] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.453034] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.453375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.453406] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.453738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.453779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.454006] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.454036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.454248] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.454278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.449 [2024-11-06 09:08:59.454533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.449 [2024-11-06 09:08:59.454566] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.449 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.454822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.454853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.455066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.455098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.455471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.455502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.455878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.455912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.456247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.456278] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.456666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.456698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.457067] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.457100] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.457437] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.457468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.457845] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.457879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.458261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.458292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.458679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.458710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.459095] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.459129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.459483] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.459514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.459868] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.459902] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.460290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.460321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.460680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.460711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.461106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.461140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.461480] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.461512] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.461885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.461917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.462287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.462319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.462657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.462689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.463059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.463092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.463438] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.463471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.463876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.463910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.464165] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.464196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.464522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.464554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.464940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.464974] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.465227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.465260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.465506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.465538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.465807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.465841] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.466066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.466099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.466471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.466503] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.466876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.466909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.467303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.467334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.467706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.467744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.468119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.468150] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.468516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.468547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.468893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.468926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.469282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.469314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.469663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.469694] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.470055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.470088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.470424] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.470457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.470782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.470817] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.471189] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.471220] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.471591] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.471623] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.471976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.472008] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.472385] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.472417] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.472629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.472662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.473070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.473103] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.473428] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.473461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.473849] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.473883] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.474256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.474289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.474655] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.474687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.475068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.475101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.475471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.475502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.475876] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.475909] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.476290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.476322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.450 qpair failed and we were unable to recover it. 00:33:09.450 [2024-11-06 09:08:59.476571] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.450 [2024-11-06 09:08:59.476603] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.476944] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.476978] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.477368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.477400] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.477651] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.477682] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.478064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.478099] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.478316] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.478349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.478725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.478766] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.479145] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.479175] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.479500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.479531] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.479890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.479923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.480251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.480281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.480657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.480689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.481073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.481106] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.481487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.481518] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.481899] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.481932] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.482283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.482314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.482661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.482695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.483072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.483113] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.483471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.483502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.483881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.483914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.484290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.484321] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.484675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.484706] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.485096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.485129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.485472] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.485504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.485859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.485893] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.486238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.486270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.486518] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.486550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.486885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.486918] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.487321] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.487352] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.487684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.487717] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.488094] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.488126] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.488501] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.488532] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.488864] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.488898] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.489129] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.489161] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.489487] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.489519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.489907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.489940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.490295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.490326] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.490707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.490738] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.491086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.491118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.491485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.491517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.491894] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.491926] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.492300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.492331] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.492702] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.492735] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.493151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.493183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.493560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.493591] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.493851] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.493884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.494245] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.494277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.494656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.494687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.495055] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.495088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.495418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.495448] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.495834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.451 [2024-11-06 09:08:59.495866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.451 qpair failed and we were unable to recover it. 00:33:09.451 [2024-11-06 09:08:59.496261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.496292] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.496675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.496707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.497035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.497068] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.497470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.497501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.497873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.497905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.498244] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.498275] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.498629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.498667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.498907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.498940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.499262] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.499293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.499650] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.499681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.500077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.500109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.500494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.500524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.500883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.500916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.501290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.501320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.501698] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.501730] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.502102] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.502134] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.502466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.502497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.502771] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.502807] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.503176] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.503207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.503545] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.503577] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.503834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.503868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.504223] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.504254] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.504581] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.504612] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.504952] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.504984] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.505375] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.505408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.505658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.505688] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.506033] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.506065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.506372] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.506404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.506791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.506823] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.507208] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.507239] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.507602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.507634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.507963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.507996] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.508349] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.508380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.508738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.508778] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.509008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.509040] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.509420] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.509451] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.509659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.509690] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.510071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.510104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.510478] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.510510] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.510860] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.510892] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.511287] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.511318] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.511641] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.511673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.512021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.512052] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.512427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.512459] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.512835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.512866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.513106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.513136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.513470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.513508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.513875] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.513907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.514238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.514270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.514645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.514674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.514934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.514966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.515323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.515355] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.515567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.515599] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.515974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.516007] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.516368] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.516399] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.516713] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.516743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.517078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.517110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.452 qpair failed and we were unable to recover it. 00:33:09.452 [2024-11-06 09:08:59.517476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.452 [2024-11-06 09:08:59.517507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.517783] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.517816] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.518195] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.518227] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.518567] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.518600] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.518932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.518965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.519212] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.519243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.519599] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.519630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.519974] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.520005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.520387] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.520420] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.520790] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.520822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.520916] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.520944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.521289] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.521319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.521532] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.521563] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.521934] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.521965] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.522192] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.522222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.522578] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.522611] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.522827] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.522860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.523238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.523270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.523494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.523525] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.523940] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.523971] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.524352] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.524386] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.524765] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.524800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.525059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.525091] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.525469] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.525501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.525716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.525756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.526164] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.526196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.526573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.526604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.526978] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.527011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.527237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.527270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.527595] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.527633] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.528005] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.528038] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.528413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.528468] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.528672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.528703] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.529054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.529087] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.529411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.529442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.529679] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.529711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.530104] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.530137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.530466] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.530497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.530707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.530739] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.531162] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.531194] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.531533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.531565] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.531955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.531988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.532340] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.532371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.532707] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.532740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.533090] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.533121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.533413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.533444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.533684] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.533714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.533823] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.533854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.534092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.534123] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.534413] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.534444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.534832] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.534865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.535252] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.535283] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.535496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.535527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.535781] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.535815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.536065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.536096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.536473] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.536506] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.536773] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.536805] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.453 [2024-11-06 09:08:59.537152] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.453 [2024-11-06 09:08:59.537184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.453 qpair failed and we were unable to recover it. 00:33:09.728 [2024-11-06 09:08:59.537531] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.728 [2024-11-06 09:08:59.537567] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.728 qpair failed and we were unable to recover it. 00:33:09.728 [2024-11-06 09:08:59.537793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.728 [2024-11-06 09:08:59.537826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.728 qpair failed and we were unable to recover it. 00:33:09.728 [2024-11-06 09:08:59.538061] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.728 [2024-11-06 09:08:59.538094] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.728 qpair failed and we were unable to recover it. 00:33:09.728 [2024-11-06 09:08:59.538412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.728 [2024-11-06 09:08:59.538442] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.728 qpair failed and we were unable to recover it. 00:33:09.728 [2024-11-06 09:08:59.538823] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.728 [2024-11-06 09:08:59.538855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.728 qpair failed and we were unable to recover it. 00:33:09.728 [2024-11-06 09:08:59.539220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.539252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.539588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.539618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.539963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.539997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.540359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.540393] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.540723] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.540762] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.541096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.541129] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.541233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.541270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.541606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.541638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.541890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.541922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.542147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.542179] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.542403] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.542436] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.542811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.542843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.543225] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.543256] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.543476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.543507] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.543628] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.543662] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.543882] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.543916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.544308] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.544340] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.544558] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.544590] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.544943] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.544976] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.545311] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.545341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.545720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.545759] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.546116] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.546149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.546526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.729 [2024-11-06 09:08:59.546559] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.729 qpair failed and we were unable to recover it. 00:33:09.729 [2024-11-06 09:08:59.546915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.546948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.547182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.547217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.547538] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.547569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.547795] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.547827] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.548206] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.548238] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.548625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.548656] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.548893] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.548925] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.549282] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.549316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.549648] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.549680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.550064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.550097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.550315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.550349] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.550715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.550756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.551117] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.551149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.551506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.551537] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.551769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.551800] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.552037] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.552069] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.552304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.552334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.552569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.552601] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.552938] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.552970] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.553313] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.553346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.553449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.553480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.553883] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.553916] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.554170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.554201] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.554449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.554487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.554712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.554743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.730 [2024-11-06 09:08:59.555115] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.730 [2024-11-06 09:08:59.555146] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.730 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.555548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.555580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.555949] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.555983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.556338] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.556369] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.556742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.556784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.557186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.557217] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.557582] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.557613] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.557835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.557868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.558266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.558298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.558680] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.558711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.559054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.559088] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.559417] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.559449] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.559811] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.559843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.560100] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.560133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.560510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.560546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.560910] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.560943] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.561346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.561380] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.561780] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.561815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.562190] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.562221] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.562552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.562584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.562955] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.562989] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.563348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.563381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.563760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.563793] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.564148] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.564180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.564536] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.564569] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.731 qpair failed and we were unable to recover it. 00:33:09.731 [2024-11-06 09:08:59.564908] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.731 [2024-11-06 09:08:59.564941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.565328] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.565361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.565712] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.565777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.566021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.566055] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.566434] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.566465] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.566825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.566859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.567227] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.567258] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.567588] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.567622] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.567976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.568009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.568333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.568365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.568717] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.568758] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.569134] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.569167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.569541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.569574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.569906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.569947] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.570325] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.570357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.570729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.570770] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.571071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.571104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.571462] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.571494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.571878] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.571912] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.572271] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.572306] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.572677] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.572709] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.572947] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.572980] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.573362] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.573394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.573743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.573784] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.574151] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.574185] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.732 qpair failed and we were unable to recover it. 00:33:09.732 [2024-11-06 09:08:59.574552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.732 [2024-11-06 09:08:59.574584] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.574932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.574963] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.575346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.575378] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.575602] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.575634] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.575881] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.575915] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.576285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.576316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.576618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.576651] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.577019] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.577054] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.577295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.577327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.577710] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.577741] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.578103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.578136] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.578515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.578550] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.578906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.578939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.579310] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.579345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.579734] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.579788] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.580185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.580218] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.580537] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.580570] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.580897] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.580931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.581326] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.581358] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.581716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.581755] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.582125] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.582156] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.582530] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.582564] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.582933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.733 [2024-11-06 09:08:59.582964] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.733 qpair failed and we were unable to recover it. 00:33:09.733 [2024-11-06 09:08:59.583322] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.583353] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.583685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.583718] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.584002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.584036] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.584412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.584444] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.584777] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.584811] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.585158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.585196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.585520] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.585551] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.585869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.585903] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.586283] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.586316] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.586656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.586687] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.587064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.587097] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.587450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.587482] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.587823] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.587855] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.588249] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.588281] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.588634] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.588667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.589033] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.589065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.589439] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.589470] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.589825] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.589858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.590229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.590262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.590640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.590671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.590912] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.590945] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.591301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.591332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.591708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.591740] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.592135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.592167] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.592496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.592527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.734 [2024-11-06 09:08:59.592902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.734 [2024-11-06 09:08:59.592935] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.734 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.593290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.593323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.593696] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.593727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.594078] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.594110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.594485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.594517] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.594835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.594868] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.595218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.595251] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.595605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.595638] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.595977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.596009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.596388] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.596419] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.596775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.596808] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.597128] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.597159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.597416] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.597447] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.597810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.597843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.598173] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.598206] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.598457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.598489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.598833] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.598865] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.599258] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.735 [2024-11-06 09:08:59.599291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.735 qpair failed and we were unable to recover it. 00:33:09.735 [2024-11-06 09:08:59.599657] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.599689] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.600077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.600109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.600506] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.600544] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.600886] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.600920] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.601262] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.601294] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.601687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.601719] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.602077] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.602110] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.602361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.602395] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.602613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.602646] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.602826] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.602857] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.603280] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.603312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.603649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.603681] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.603909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.603942] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.604291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.604323] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.604738] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.604781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.605183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.605214] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.605573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.605605] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.605843] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.605880] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.606256] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.606287] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.606637] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.606670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.607012] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.607044] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.736 [2024-11-06 09:08:59.607262] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.736 [2024-11-06 09:08:59.607295] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.736 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.607660] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.607692] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.608071] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.608105] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.608474] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.608504] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.608909] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.608940] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.609096] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.609128] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.609354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.609384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.609729] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.609769] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.610150] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.610183] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.610409] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.610439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.610733] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.610773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.611118] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.611149] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.611515] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.611546] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.611791] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.611826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.612179] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.612210] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.612449] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.612479] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.612782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.612815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.613172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.613204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.613555] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.613585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.613828] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.613860] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.613966] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.613994] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.614498] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.614572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.614910] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.737 [2024-11-06 09:08:59.614928] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.737 qpair failed and we were unable to recover it. 00:33:09.737 [2024-11-06 09:08:59.615396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.615461] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.615892] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.615911] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.616106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.616121] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.616496] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.616508] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.616709] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.616722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.617188] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.617253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.617500] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.617516] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.617740] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.617764] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.618088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.618153] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.618292] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.618308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.618502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.618514] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.618697] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.618710] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.619257] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.619367] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.620032] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.620142] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.620546] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.620585] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.620800] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.620834] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.621229] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.621261] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.621629] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.621661] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.621954] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.621987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.622396] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.622428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.622793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.622825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.623106] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.623137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.623348] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.623381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.623667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.623698] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.624123] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.624157] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.738 qpair failed and we were unable to recover it. 00:33:09.738 [2024-11-06 09:08:59.624533] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.738 [2024-11-06 09:08:59.624576] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.624810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.624843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.625247] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.625280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.625379] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.625410] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.625810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.625843] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.626226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.626257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.626627] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.626657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.627023] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.627057] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.627432] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.627464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.627585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.627616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.627980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.628012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.628376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.628407] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.628731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.628773] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.629107] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.629137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.629364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.629394] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.629645] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.629676] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.630043] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.630076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.630450] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.630480] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.630818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.630850] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.631239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.631270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.631484] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.631515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.631611] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.631639] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.631900] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.631930] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.632133] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.632162] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.632597] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.632627] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.632981] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.633012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.739 [2024-11-06 09:08:59.633394] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.739 [2024-11-06 09:08:59.633425] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.739 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.633639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.633671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.634089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.634122] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.634504] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.634535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.634806] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.634859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.635113] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.635144] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.635529] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.635560] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.635979] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.636010] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.636363] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.636396] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.636716] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.636763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.637131] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.637163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.637516] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.637547] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.637873] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.637905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.638149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.638180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.638534] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.638572] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.638782] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.638814] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.639191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.639222] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.639443] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.639474] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.639816] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.639849] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.640253] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.640284] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.640640] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.740 [2024-11-06 09:08:59.640671] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.740 qpair failed and we were unable to recover it. 00:33:09.740 [2024-11-06 09:08:59.640930] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.744 [2024-11-06 09:08:59.640962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.744 qpair failed and we were unable to recover it. 00:33:09.744 [2024-11-06 09:08:59.641064] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.744 [2024-11-06 09:08:59.641092] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.744 qpair failed and we were unable to recover it. 00:33:09.744 [2024-11-06 09:08:59.641459] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.744 [2024-11-06 09:08:59.641489] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.744 qpair failed and we were unable to recover it. 00:33:09.744 [2024-11-06 09:08:59.641856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.744 [2024-11-06 09:08:59.641888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.744 qpair failed and we were unable to recover it. 00:33:09.744 [2024-11-06 09:08:59.642267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.642298] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.642522] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.642553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.642842] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.642875] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.643231] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.643260] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.643494] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.643526] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:09.745 [2024-11-06 09:08:59.643890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.643923] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@864 -- # return 0 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.644262] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:33:09.745 [2024-11-06 09:08:59.644293] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:09.745 [2024-11-06 09:08:59.644661] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.644693] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:09.745 [2024-11-06 09:08:59.645038] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.645070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.645448] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.645487] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.645841] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.645874] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.646108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.646137] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.646544] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.646573] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.646796] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.646826] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.647215] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.647245] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.647471] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.647502] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.647907] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.647937] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.648300] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.648332] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.648693] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.648722] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.648976] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.649005] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.649259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.649289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.649664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.649695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.650066] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.650098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.650315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.650344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.745 [2024-11-06 09:08:59.650711] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.745 [2024-11-06 09:08:59.650743] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.745 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.651012] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.651041] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.651290] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.651319] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.651569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.651604] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.651853] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.651884] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.652228] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.652257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.652604] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.652632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.653016] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.653046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.653295] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.653330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.653613] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.653644] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.653915] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.653948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.654315] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.654347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.654706] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.654736] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.655160] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.655191] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.655410] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.655439] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.655775] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.655806] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.656185] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.656216] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.656454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.656483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.656885] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.656917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.657154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.657184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.657600] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.657629] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.657856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.657887] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.746 [2024-11-06 09:08:59.658138] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.746 [2024-11-06 09:08:59.658168] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.746 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.658414] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.658446] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.658836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.658866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.659103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.659132] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.659378] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.659408] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.659797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.659830] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.660273] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.660303] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.660668] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.660697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.661089] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.661120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.661554] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.661582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.661871] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.661901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.662141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.662173] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.662521] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.662549] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.662821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.662854] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.663291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.663320] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.663685] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.663714] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.663982] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.664012] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.664406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.664435] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.664835] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.664866] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.665198] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.665228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.665476] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.665505] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.665732] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.665799] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.666081] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.666140] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.747 [2024-11-06 09:08:59.666383] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.747 [2024-11-06 09:08:59.666414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.747 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.666807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.666840] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.667209] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.667240] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.667589] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.667618] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.668025] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.668056] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.668465] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.668494] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.668857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.668888] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.669103] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.669133] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.669376] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.669404] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.669797] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.669829] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.670056] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.670086] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.670404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.670434] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.670793] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.670825] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.671172] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.671200] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.671624] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.671654] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.671863] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.671895] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.672261] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.672291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.672667] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.672697] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.673035] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.673065] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.673293] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.673322] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.673715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.673744] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.674015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.674045] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.674267] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.674296] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.674552] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.674582] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.674813] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.674846] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.675183] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.675215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.675560] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.675589] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.675854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.675885] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.676124] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.676152] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.748 [2024-11-06 09:08:59.676509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.748 [2024-11-06 09:08:59.676539] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.748 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.676932] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.676962] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.677332] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.677361] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.677735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.677775] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.678134] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.678163] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.678542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.678571] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7f6088000b90 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.678810] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.678862] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.679191] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.679204] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.679433] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.679454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.679751] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.679772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.680205] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.680267] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.680621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.680636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.680695] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.680705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.681186] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.681250] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.681467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.681481] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.681840] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.681853] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.682170] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.682181] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.682389] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.682402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.682757] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.682772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.683147] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.683160] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.683460] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.683471] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.683809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.683822] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.684141] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.684154] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.684526] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.684538] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.684769] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.684781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.685143] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.685155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.749 [2024-11-06 09:08:59.685541] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.749 [2024-11-06 09:08:59.685553] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.749 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.685855] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.685867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.686182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.686193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.686542] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.686554] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.686867] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.686877] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.687239] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.687253] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.687609] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.687620] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.687970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.687981] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.688180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.688193] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:33:09.750 [2024-11-06 09:08:59.688574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.688588] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:33:09.750 [2024-11-06 09:08:59.688891] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.688905] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.750 [2024-11-06 09:08:59.689243] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.689257] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:09.750 [2024-11-06 09:08:59.689470] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.689483] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.689809] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.689821] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.690158] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.690170] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.690510] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.690522] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.690857] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.690870] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.691196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.691207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.691369] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.691381] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.691720] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.691732] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.692092] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.692104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.692406] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.692416] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.692756] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.692772] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.693109] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.693120] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.693454] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.750 [2024-11-06 09:08:59.693464] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.750 qpair failed and we were unable to recover it. 00:33:09.750 [2024-11-06 09:08:59.693770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.693783] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.694085] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.694096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.694334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.694344] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.694649] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.694659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.694980] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.694991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.695327] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.695337] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.695666] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.695677] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.696008] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.696019] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.696304] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.696315] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.696511] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.696521] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.696848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.696859] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.697204] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.697215] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.697523] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.697535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.697898] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.697910] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.698265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.698276] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.698461] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.698473] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.698821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.698832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.699168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.699178] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.699525] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.699535] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.699896] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.699907] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.700238] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.700248] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.700585] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.700595] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.700927] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.700939] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.701296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.701307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.701669] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.701680] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.702004] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.702014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.702317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.702327] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.702638] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.751 [2024-11-06 09:08:59.702648] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.751 qpair failed and we were unable to recover it. 00:33:09.751 [2024-11-06 09:08:59.702989] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.703000] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.703346] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.703357] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.703658] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.703668] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.704021] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.704031] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.704344] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.704356] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.704683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.704695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.705059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.705071] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.705429] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.705440] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.705770] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.705781] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.706105] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.706115] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.706418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.706428] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.706766] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.706777] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.707097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.707107] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.707447] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.707457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.707807] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.707819] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.708154] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.708164] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.708467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.708477] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.708822] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.708833] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.709149] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.709159] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.709509] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.709519] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.709869] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.709879] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.710199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.710209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.710513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.710523] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.710856] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.710867] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.752 [2024-11-06 09:08:59.711119] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.752 [2024-11-06 09:08:59.711130] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.752 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.711333] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.711345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.711675] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.711686] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.712018] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.712028] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.712358] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.712368] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.712664] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.712675] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.712986] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.712997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.713334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.713345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.713693] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.713705] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.714041] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.714053] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.714392] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.714402] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.714742] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.714756] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.715097] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.715108] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.715285] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.715299] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.715644] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.715655] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.715969] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.715979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.716319] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.716330] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.716663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.716673] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.716998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.717009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.717339] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.717351] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.717672] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.717683] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.718015] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.718026] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.718355] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.718365] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.718573] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.718583] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.753 qpair failed and we were unable to recover it. 00:33:09.753 [2024-11-06 09:08:59.718911] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.753 [2024-11-06 09:08:59.718922] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.719135] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.719145] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.719226] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.719237] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.719401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.719411] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.719735] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.719745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.720086] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.720096] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.720402] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.720412] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.720753] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.720763] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.720981] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.720991] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.721281] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.721291] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.721479] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.721490] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.721836] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.721847] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.722180] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.722190] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.722374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.722385] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.722737] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.722751] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.723054] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.723064] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.723237] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.723247] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 Malloc0 00:33:09.754 [2024-11-06 09:08:59.723647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.723659] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.723848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.723861] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.724087] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.724098] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.724427] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.724438] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.754 [2024-11-06 09:08:59.724642] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.724653] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.724970] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.724983] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.754 [2024-11-06 09:08:59.725334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.725346] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:09.754 [2024-11-06 09:08:59.725695] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.754 [2024-11-06 09:08:59.725707] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.754 qpair failed and we were unable to recover it. 00:33:09.754 [2024-11-06 09:08:59.726059] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.726070] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.726374] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.726384] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.726694] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.726704] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.726973] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.726987] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.727199] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.727209] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.727467] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.727478] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.727804] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.727815] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.728130] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.728141] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.728354] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.728363] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.728686] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.728696] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.728906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.728917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.729242] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.729252] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.729457] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.729466] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.729518] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.729527] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.729848] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.729858] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.730070] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.730081] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.730412] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.730422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.730619] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.730630] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.730942] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.730953] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.731144] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.731155] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.731162] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:33:09.755 [2024-11-06 09:08:59.731359] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.731371] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.731731] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.731745] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.732051] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.732062] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.755 [2024-11-06 09:08:59.732233] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.755 [2024-11-06 09:08:59.732243] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.755 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.732551] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.732561] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.732906] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.732917] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.733218] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.733228] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.733444] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.733454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.733743] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.733760] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.733963] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.733973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.734194] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.734212] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.734626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.734636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.735017] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.735029] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.735331] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.735341] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.735547] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.735558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.735865] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.735876] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.736182] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.736192] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.736411] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.736422] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.736606] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.736615] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.736834] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.736845] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.737220] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.737230] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.737548] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.737558] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.737859] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.737869] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.738093] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.738104] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.738446] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.738457] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.738646] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.738658] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.738929] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.738941] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.739101] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.739111] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.739505] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.739515] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.739708] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.739720] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.739920] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.739931] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 [2024-11-06 09:08:59.740263] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.740274] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.756 qpair failed and we were unable to recover it. 00:33:09.756 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.756 [2024-11-06 09:08:59.740485] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.756 [2024-11-06 09:08:59.740497] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:33:09.757 [2024-11-06 09:08:59.740818] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.740831] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.757 [2024-11-06 09:08:59.741168] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.741180] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:09.757 [2024-11-06 09:08:59.741513] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.741524] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.741725] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.741737] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.742079] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.742089] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.742405] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.742415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.742767] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.742779] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.743099] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.743109] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.743291] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.743301] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.743701] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.743711] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.744108] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.744118] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.744451] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.744462] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.744821] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.744832] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.745174] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.745184] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.745486] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.745496] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.745854] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.745864] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.746065] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.746076] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.746303] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.746314] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.746445] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.746454] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.746647] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.746657] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.746890] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.746901] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.747196] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.747207] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.747491] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.747501] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.747691] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.747702] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.757 [2024-11-06 09:08:59.748040] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.757 [2024-11-06 09:08:59.748051] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.757 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.748381] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.748392] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.748569] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.748580] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.748923] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.748934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.748998] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.749009] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.749184] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.749196] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.749404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.749415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.749722] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.749733] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.749957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.749969] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.750298] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.750309] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.750612] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.750624] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.750985] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.750997] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.751336] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.751347] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.751663] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.751674] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.752002] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.752014] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.752364] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.758 [2024-11-06 09:08:59.752376] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.752730] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.752742] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:33:09.758 [2024-11-06 09:08:59.753068] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.753080] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.758 [2024-11-06 09:08:59.753401] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.753414] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:09.758 [2024-11-06 09:08:59.753749] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.753761] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.753956] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.753966] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.754301] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.754312] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.754618] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.754628] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.755005] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.755016] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.755343] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.755354] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.758 [2024-11-06 09:08:59.755721] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.758 [2024-11-06 09:08:59.755731] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.758 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.756073] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.756085] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.756268] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.756280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.756626] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.756637] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.756990] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.757001] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.757323] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.757334] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.757683] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.757695] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.758034] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.758046] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.758269] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.758280] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.758605] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.758616] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.758945] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.758957] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.759251] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.759262] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.759614] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.759625] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.759933] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.759944] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.760265] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.760277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.760574] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.760586] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.760936] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.760948] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.761266] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.761277] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.761625] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.761636] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.761977] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.761988] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.762334] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.762345] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.762703] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.762715] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.762902] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.762914] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.763259] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.763270] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.763563] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.763574] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 [2024-11-06 09:08:59.763922] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.763934] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.759 qpair failed and we were unable to recover it. 00:33:09.759 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.759 [2024-11-06 09:08:59.764277] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.759 [2024-11-06 09:08:59.764289] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.764639] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.764650] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.764961] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.764973] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.760 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:09.760 [2024-11-06 09:08:59.765317] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.765328] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.765621] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.765632] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.765957] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.765972] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.766296] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.766307] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.766656] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.766667] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.766999] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.767011] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.767361] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.767372] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.767687] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.767699] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.768048] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.768059] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.768398] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.760 [2024-11-06 09:08:59.768409] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.760 qpair failed and we were unable to recover it. 00:33:09.760 [2024-11-06 09:08:59.768760] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.764 [2024-11-06 09:08:59.768771] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.764 qpair failed and we were unable to recover it. 00:33:09.764 [2024-11-06 09:08:59.769072] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.764 [2024-11-06 09:08:59.769084] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.764 qpair failed and we were unable to recover it. 00:33:09.764 [2024-11-06 09:08:59.769418] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.764 [2024-11-06 09:08:59.769429] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.764 qpair failed and we were unable to recover it. 00:33:09.764 [2024-11-06 09:08:59.769659] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.764 [2024-11-06 09:08:59.769670] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.764 qpair failed and we were unable to recover it. 00:33:09.764 [2024-11-06 09:08:59.769968] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.764 [2024-11-06 09:08:59.769979] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.764 qpair failed and we were unable to recover it. 00:33:09.764 [2024-11-06 09:08:59.770297] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.764 [2024-11-06 09:08:59.770308] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.765 qpair failed and we were unable to recover it. 00:33:09.765 [2024-11-06 09:08:59.770502] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.765 [2024-11-06 09:08:59.770513] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.765 qpair failed and we were unable to recover it. 00:33:09.765 [2024-11-06 09:08:59.770715] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.765 [2024-11-06 09:08:59.770727] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.765 qpair failed and we were unable to recover it. 00:33:09.765 [2024-11-06 09:08:59.771088] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.765 [2024-11-06 09:08:59.771101] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.765 qpair failed and we were unable to recover it. 00:33:09.765 [2024-11-06 09:08:59.771404] posix.c:1055:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:33:09.765 [2024-11-06 09:08:59.771415] nvme_tcp.c:2288:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x8500c0 with addr=10.0.0.2, port=4420 00:33:09.765 qpair failed and we were unable to recover it. 00:33:09.765 [2024-11-06 09:08:59.771536] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:33:09.765 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.765 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:33:09.765 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:09.765 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:09.765 [2024-11-06 09:08:59.782348] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:09.765 [2024-11-06 09:08:59.782434] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:09.765 [2024-11-06 09:08:59.782458] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:09.765 [2024-11-06 09:08:59.782466] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:09.765 [2024-11-06 09:08:59.782474] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:09.765 [2024-11-06 09:08:59.782497] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:09.765 qpair failed and we were unable to recover it. 00:33:09.765 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:09.765 09:08:59 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@50 -- # wait 2080373 00:33:09.765 [2024-11-06 09:08:59.792232] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:09.765 [2024-11-06 09:08:59.792304] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:09.765 [2024-11-06 09:08:59.792321] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:09.765 [2024-11-06 09:08:59.792328] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:09.765 [2024-11-06 09:08:59.792335] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:09.765 [2024-11-06 09:08:59.792351] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:09.765 qpair failed and we were unable to recover it. 00:33:09.765 [2024-11-06 09:08:59.802229] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:09.765 [2024-11-06 09:08:59.802298] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:09.765 [2024-11-06 09:08:59.802315] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:09.765 [2024-11-06 09:08:59.802322] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:09.765 [2024-11-06 09:08:59.802329] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:09.765 [2024-11-06 09:08:59.802345] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:09.765 qpair failed and we were unable to recover it. 00:33:09.765 [2024-11-06 09:08:59.812242] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:09.765 [2024-11-06 09:08:59.812312] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:09.765 [2024-11-06 09:08:59.812328] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:09.765 [2024-11-06 09:08:59.812335] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:09.765 [2024-11-06 09:08:59.812342] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:09.765 [2024-11-06 09:08:59.812357] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:09.765 qpair failed and we were unable to recover it. 00:33:09.765 [2024-11-06 09:08:59.822205] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:09.765 [2024-11-06 09:08:59.822274] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:09.765 [2024-11-06 09:08:59.822289] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:09.765 [2024-11-06 09:08:59.822296] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:09.765 [2024-11-06 09:08:59.822302] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:09.765 [2024-11-06 09:08:59.822318] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:09.765 qpair failed and we were unable to recover it. 00:33:10.027 [2024-11-06 09:08:59.832190] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.027 [2024-11-06 09:08:59.832250] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.027 [2024-11-06 09:08:59.832267] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.027 [2024-11-06 09:08:59.832274] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.027 [2024-11-06 09:08:59.832280] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.027 [2024-11-06 09:08:59.832296] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.027 qpair failed and we were unable to recover it. 00:33:10.027 [2024-11-06 09:08:59.842126] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.027 [2024-11-06 09:08:59.842221] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.027 [2024-11-06 09:08:59.842242] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.027 [2024-11-06 09:08:59.842249] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.027 [2024-11-06 09:08:59.842256] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.027 [2024-11-06 09:08:59.842271] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.027 qpair failed and we were unable to recover it. 00:33:10.027 [2024-11-06 09:08:59.852224] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.027 [2024-11-06 09:08:59.852290] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.027 [2024-11-06 09:08:59.852304] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.027 [2024-11-06 09:08:59.852312] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.027 [2024-11-06 09:08:59.852318] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.027 [2024-11-06 09:08:59.852333] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.027 qpair failed and we were unable to recover it. 00:33:10.027 [2024-11-06 09:08:59.862324] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.027 [2024-11-06 09:08:59.862385] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.027 [2024-11-06 09:08:59.862400] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.027 [2024-11-06 09:08:59.862407] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.027 [2024-11-06 09:08:59.862413] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.027 [2024-11-06 09:08:59.862427] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.027 qpair failed and we were unable to recover it. 00:33:10.027 [2024-11-06 09:08:59.872362] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.027 [2024-11-06 09:08:59.872422] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.027 [2024-11-06 09:08:59.872437] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.027 [2024-11-06 09:08:59.872444] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.027 [2024-11-06 09:08:59.872450] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.027 [2024-11-06 09:08:59.872465] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.027 qpair failed and we were unable to recover it. 00:33:10.027 [2024-11-06 09:08:59.882343] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.027 [2024-11-06 09:08:59.882397] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.027 [2024-11-06 09:08:59.882412] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.027 [2024-11-06 09:08:59.882419] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.027 [2024-11-06 09:08:59.882425] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.027 [2024-11-06 09:08:59.882444] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.027 qpair failed and we were unable to recover it. 00:33:10.027 [2024-11-06 09:08:59.892340] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.892400] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.892414] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.892421] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.892427] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.892442] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:08:59.902391] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.902450] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.902464] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.902471] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.902477] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.902491] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:08:59.912384] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.912442] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.912457] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.912464] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.912470] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.912485] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:08:59.922421] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.922508] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.922536] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.922545] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.922552] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.922571] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:08:59.932424] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.932513] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.932540] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.932549] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.932556] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.932575] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:08:59.942487] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.942551] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.942578] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.942587] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.942594] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.942613] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:08:59.952396] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.952497] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.952513] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.952520] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.952527] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.952542] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:08:59.962513] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.962569] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.962584] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.962591] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.962597] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.962612] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:08:59.972581] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.972642] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.972661] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.972668] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.972674] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.972688] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:08:59.982587] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.982642] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.982656] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.982663] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.982669] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.982683] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:08:59.992773] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:08:59.992875] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:08:59.992891] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:08:59.992898] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:08:59.992904] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:08:59.992918] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:09:00.002624] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:09:00.002678] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:09:00.002692] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:09:00.002699] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:09:00.002706] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:09:00.002720] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:09:00.012622] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:09:00.012688] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:09:00.012704] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:09:00.012711] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:09:00.012717] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:09:00.012736] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:09:00.022643] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:09:00.022702] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:09:00.022715] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:09:00.022723] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:09:00.022729] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:09:00.022743] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:09:00.032610] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:09:00.032671] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:09:00.032685] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:09:00.032693] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.028 [2024-11-06 09:09:00.032700] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.028 [2024-11-06 09:09:00.032714] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.028 qpair failed and we were unable to recover it. 00:33:10.028 [2024-11-06 09:09:00.042769] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.028 [2024-11-06 09:09:00.042825] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.028 [2024-11-06 09:09:00.042839] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.028 [2024-11-06 09:09:00.042846] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.029 [2024-11-06 09:09:00.042852] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.029 [2024-11-06 09:09:00.042867] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.029 qpair failed and we were unable to recover it. 00:33:10.029 [2024-11-06 09:09:00.052785] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.029 [2024-11-06 09:09:00.052846] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.029 [2024-11-06 09:09:00.052860] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.029 [2024-11-06 09:09:00.052867] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.029 [2024-11-06 09:09:00.052873] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.029 [2024-11-06 09:09:00.052888] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.029 qpair failed and we were unable to recover it. 00:33:10.029 [2024-11-06 09:09:00.062821] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.029 [2024-11-06 09:09:00.062877] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.029 [2024-11-06 09:09:00.062891] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.029 [2024-11-06 09:09:00.062899] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.029 [2024-11-06 09:09:00.062905] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.029 [2024-11-06 09:09:00.062919] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.029 qpair failed and we were unable to recover it. 00:33:10.029 [2024-11-06 09:09:00.072897] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.029 [2024-11-06 09:09:00.072950] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.029 [2024-11-06 09:09:00.072964] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.029 [2024-11-06 09:09:00.072971] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.029 [2024-11-06 09:09:00.072977] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.029 [2024-11-06 09:09:00.072991] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.029 qpair failed and we were unable to recover it. 00:33:10.029 [2024-11-06 09:09:00.082876] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.029 [2024-11-06 09:09:00.082973] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.029 [2024-11-06 09:09:00.082988] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.029 [2024-11-06 09:09:00.082995] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.029 [2024-11-06 09:09:00.083001] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.029 [2024-11-06 09:09:00.083016] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.029 qpair failed and we were unable to recover it. 00:33:10.029 [2024-11-06 09:09:00.092906] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.029 [2024-11-06 09:09:00.092965] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.029 [2024-11-06 09:09:00.092979] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.029 [2024-11-06 09:09:00.092986] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.029 [2024-11-06 09:09:00.092993] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.029 [2024-11-06 09:09:00.093007] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.029 qpair failed and we were unable to recover it. 00:33:10.029 [2024-11-06 09:09:00.102952] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.029 [2024-11-06 09:09:00.103003] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.029 [2024-11-06 09:09:00.103020] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.029 [2024-11-06 09:09:00.103027] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.029 [2024-11-06 09:09:00.103033] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.029 [2024-11-06 09:09:00.103048] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.029 qpair failed and we were unable to recover it. 00:33:10.029 [2024-11-06 09:09:00.112959] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.029 [2024-11-06 09:09:00.113011] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.029 [2024-11-06 09:09:00.113024] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.029 [2024-11-06 09:09:00.113031] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.029 [2024-11-06 09:09:00.113038] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.029 [2024-11-06 09:09:00.113051] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.029 qpair failed and we were unable to recover it. 00:33:10.029 [2024-11-06 09:09:00.122989] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.029 [2024-11-06 09:09:00.123047] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.029 [2024-11-06 09:09:00.123061] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.029 [2024-11-06 09:09:00.123069] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.029 [2024-11-06 09:09:00.123075] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.029 [2024-11-06 09:09:00.123088] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.029 qpair failed and we were unable to recover it. 00:33:10.029 [2024-11-06 09:09:00.133050] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.029 [2024-11-06 09:09:00.133148] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.029 [2024-11-06 09:09:00.133162] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.029 [2024-11-06 09:09:00.133170] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.029 [2024-11-06 09:09:00.133176] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.029 [2024-11-06 09:09:00.133190] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.029 qpair failed and we were unable to recover it. 00:33:10.292 [2024-11-06 09:09:00.143058] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.292 [2024-11-06 09:09:00.143114] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.292 [2024-11-06 09:09:00.143127] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.292 [2024-11-06 09:09:00.143134] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.292 [2024-11-06 09:09:00.143141] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.292 [2024-11-06 09:09:00.143158] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.292 qpair failed and we were unable to recover it. 00:33:10.292 [2024-11-06 09:09:00.153050] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.292 [2024-11-06 09:09:00.153103] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.292 [2024-11-06 09:09:00.153117] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.292 [2024-11-06 09:09:00.153124] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.292 [2024-11-06 09:09:00.153131] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.292 [2024-11-06 09:09:00.153145] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.292 qpair failed and we were unable to recover it. 00:33:10.292 [2024-11-06 09:09:00.163106] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.292 [2024-11-06 09:09:00.163158] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.292 [2024-11-06 09:09:00.163172] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.292 [2024-11-06 09:09:00.163179] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.292 [2024-11-06 09:09:00.163185] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.292 [2024-11-06 09:09:00.163199] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.292 qpair failed and we were unable to recover it. 00:33:10.292 [2024-11-06 09:09:00.173150] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.292 [2024-11-06 09:09:00.173207] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.292 [2024-11-06 09:09:00.173220] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.292 [2024-11-06 09:09:00.173227] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.292 [2024-11-06 09:09:00.173233] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.292 [2024-11-06 09:09:00.173247] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.292 qpair failed and we were unable to recover it. 00:33:10.292 [2024-11-06 09:09:00.183171] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.292 [2024-11-06 09:09:00.183227] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.292 [2024-11-06 09:09:00.183241] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.292 [2024-11-06 09:09:00.183248] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.292 [2024-11-06 09:09:00.183254] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.292 [2024-11-06 09:09:00.183267] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.292 qpair failed and we were unable to recover it. 00:33:10.292 [2024-11-06 09:09:00.193242] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.292 [2024-11-06 09:09:00.193294] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.292 [2024-11-06 09:09:00.193307] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.292 [2024-11-06 09:09:00.193314] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.292 [2024-11-06 09:09:00.193320] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.292 [2024-11-06 09:09:00.193334] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.292 qpair failed and we were unable to recover it. 00:33:10.292 [2024-11-06 09:09:00.203212] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.292 [2024-11-06 09:09:00.203271] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.292 [2024-11-06 09:09:00.203285] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.292 [2024-11-06 09:09:00.203292] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.292 [2024-11-06 09:09:00.203299] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.292 [2024-11-06 09:09:00.203313] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.292 qpair failed and we were unable to recover it. 00:33:10.292 [2024-11-06 09:09:00.213257] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.292 [2024-11-06 09:09:00.213315] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.292 [2024-11-06 09:09:00.213328] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.292 [2024-11-06 09:09:00.213336] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.292 [2024-11-06 09:09:00.213342] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.292 [2024-11-06 09:09:00.213356] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.292 qpair failed and we were unable to recover it. 00:33:10.292 [2024-11-06 09:09:00.223285] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.292 [2024-11-06 09:09:00.223347] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.292 [2024-11-06 09:09:00.223360] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.292 [2024-11-06 09:09:00.223367] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.292 [2024-11-06 09:09:00.223374] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.292 [2024-11-06 09:09:00.223388] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.292 qpair failed and we were unable to recover it. 00:33:10.292 [2024-11-06 09:09:00.233305] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.233359] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.233376] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.233383] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.233389] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.233403] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.243311] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.243361] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.243374] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.243382] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.243388] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.243402] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.253373] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.253432] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.253445] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.253452] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.253458] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.253472] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.263402] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.263490] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.263504] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.263511] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.263517] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.263531] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.273426] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.273492] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.273518] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.273527] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.273534] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.273557] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.283427] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.283484] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.283509] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.283518] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.283525] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.283544] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.293503] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.293564] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.293590] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.293599] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.293606] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.293625] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.303552] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.303610] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.303626] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.303633] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.303639] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.303655] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.313531] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.313583] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.313597] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.313605] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.313611] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.313625] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.323611] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.323665] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.323679] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.323686] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.323692] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.323706] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.333595] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.333654] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.333668] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.333675] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.333681] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.333695] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.343624] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.343675] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.343688] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.343695] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.343701] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.343715] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.353636] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.353684] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.353698] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.293 [2024-11-06 09:09:00.353705] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.293 [2024-11-06 09:09:00.353712] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.293 [2024-11-06 09:09:00.353725] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.293 qpair failed and we were unable to recover it. 00:33:10.293 [2024-11-06 09:09:00.363672] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.293 [2024-11-06 09:09:00.363726] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.293 [2024-11-06 09:09:00.363743] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.294 [2024-11-06 09:09:00.363756] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.294 [2024-11-06 09:09:00.363762] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.294 [2024-11-06 09:09:00.363777] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.294 qpair failed and we were unable to recover it. 00:33:10.294 [2024-11-06 09:09:00.373583] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.294 [2024-11-06 09:09:00.373644] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.294 [2024-11-06 09:09:00.373658] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.294 [2024-11-06 09:09:00.373665] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.294 [2024-11-06 09:09:00.373672] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.294 [2024-11-06 09:09:00.373685] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.294 qpair failed and we were unable to recover it. 00:33:10.294 [2024-11-06 09:09:00.383777] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.294 [2024-11-06 09:09:00.383839] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.294 [2024-11-06 09:09:00.383853] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.294 [2024-11-06 09:09:00.383861] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.294 [2024-11-06 09:09:00.383867] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.294 [2024-11-06 09:09:00.383881] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.294 qpair failed and we were unable to recover it. 00:33:10.294 [2024-11-06 09:09:00.393756] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.294 [2024-11-06 09:09:00.393849] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.294 [2024-11-06 09:09:00.393863] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.294 [2024-11-06 09:09:00.393869] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.294 [2024-11-06 09:09:00.393877] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.294 [2024-11-06 09:09:00.393890] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.294 qpair failed and we were unable to recover it. 00:33:10.556 [2024-11-06 09:09:00.403790] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.556 [2024-11-06 09:09:00.403852] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.556 [2024-11-06 09:09:00.403867] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.556 [2024-11-06 09:09:00.403874] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.403880] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.403899] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.413782] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.413843] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.413857] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.413864] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.413870] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.413884] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.423780] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.423839] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.423852] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.423859] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.423865] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.423880] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.433847] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.433900] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.433914] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.433921] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.433927] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.433941] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.444319] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.444370] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.444384] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.444391] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.444398] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.444411] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.453929] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.453988] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.454002] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.454009] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.454015] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.454029] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.463964] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.464021] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.464034] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.464041] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.464047] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.464061] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.473981] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.474033] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.474046] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.474053] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.474060] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.474073] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.483980] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.484035] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.484048] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.484055] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.484061] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.484075] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.494054] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.494111] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.494124] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.494135] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.494141] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.494154] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.504044] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.504100] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.504113] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.504120] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.504127] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.504141] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.514103] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.514155] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.514168] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.514175] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.514181] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.514195] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.524101] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.524152] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.524166] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.524173] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.557 [2024-11-06 09:09:00.524179] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.557 [2024-11-06 09:09:00.524192] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.557 qpair failed and we were unable to recover it. 00:33:10.557 [2024-11-06 09:09:00.534149] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.557 [2024-11-06 09:09:00.534209] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.557 [2024-11-06 09:09:00.534223] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.557 [2024-11-06 09:09:00.534230] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.534237] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.534254] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.544171] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.544225] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.544238] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.544245] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.544252] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.544265] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.554182] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.554273] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.554287] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.554294] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.554301] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.554314] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.564239] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.564326] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.564339] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.564346] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.564352] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.564366] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.574276] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.574335] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.574348] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.574355] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.574361] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.574375] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.584204] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.584266] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.584282] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.584289] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.584295] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.584310] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.594344] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.594399] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.594413] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.594420] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.594426] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.594440] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.604342] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.604392] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.604406] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.604413] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.604419] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.604434] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.614393] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.614455] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.614469] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.614477] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.614484] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.614498] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.624439] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.624503] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.624529] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.624542] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.624549] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.624568] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.634379] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.634440] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.634466] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.634474] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.634481] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.634500] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.644459] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.644520] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.644546] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.644554] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.644561] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.644580] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.654466] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.654523] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.654539] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.654546] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.654554] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.558 [2024-11-06 09:09:00.654569] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.558 qpair failed and we were unable to recover it. 00:33:10.558 [2024-11-06 09:09:00.664541] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.558 [2024-11-06 09:09:00.664602] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.558 [2024-11-06 09:09:00.664616] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.558 [2024-11-06 09:09:00.664623] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.558 [2024-11-06 09:09:00.664629] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.559 [2024-11-06 09:09:00.664648] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.559 qpair failed and we were unable to recover it. 00:33:10.822 [2024-11-06 09:09:00.674558] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.822 [2024-11-06 09:09:00.674610] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.822 [2024-11-06 09:09:00.674624] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.822 [2024-11-06 09:09:00.674631] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.822 [2024-11-06 09:09:00.674638] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.822 [2024-11-06 09:09:00.674652] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.822 qpair failed and we were unable to recover it. 00:33:10.822 [2024-11-06 09:09:00.684578] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.822 [2024-11-06 09:09:00.684630] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.822 [2024-11-06 09:09:00.684644] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.822 [2024-11-06 09:09:00.684651] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.822 [2024-11-06 09:09:00.684658] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.822 [2024-11-06 09:09:00.684672] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.822 qpair failed and we were unable to recover it. 00:33:10.822 [2024-11-06 09:09:00.694603] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.822 [2024-11-06 09:09:00.694657] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.822 [2024-11-06 09:09:00.694671] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.822 [2024-11-06 09:09:00.694677] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.822 [2024-11-06 09:09:00.694684] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.822 [2024-11-06 09:09:00.694698] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.822 qpair failed and we were unable to recover it. 00:33:10.822 [2024-11-06 09:09:00.704647] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.822 [2024-11-06 09:09:00.704707] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.822 [2024-11-06 09:09:00.704720] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.822 [2024-11-06 09:09:00.704727] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.822 [2024-11-06 09:09:00.704734] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.822 [2024-11-06 09:09:00.704752] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.822 qpair failed and we were unable to recover it. 00:33:10.822 [2024-11-06 09:09:00.714638] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.822 [2024-11-06 09:09:00.714736] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.822 [2024-11-06 09:09:00.714756] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.822 [2024-11-06 09:09:00.714763] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.822 [2024-11-06 09:09:00.714770] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.822 [2024-11-06 09:09:00.714784] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.822 qpair failed and we were unable to recover it. 00:33:10.822 [2024-11-06 09:09:00.724687] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.822 [2024-11-06 09:09:00.724753] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.822 [2024-11-06 09:09:00.724767] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.822 [2024-11-06 09:09:00.724774] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.822 [2024-11-06 09:09:00.724781] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.822 [2024-11-06 09:09:00.724795] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.822 qpair failed and we were unable to recover it. 00:33:10.822 [2024-11-06 09:09:00.734710] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.822 [2024-11-06 09:09:00.734770] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.822 [2024-11-06 09:09:00.734786] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.822 [2024-11-06 09:09:00.734793] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.822 [2024-11-06 09:09:00.734800] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.822 [2024-11-06 09:09:00.734815] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.822 qpair failed and we were unable to recover it. 00:33:10.822 [2024-11-06 09:09:00.744740] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.822 [2024-11-06 09:09:00.744807] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.822 [2024-11-06 09:09:00.744822] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.822 [2024-11-06 09:09:00.744829] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.822 [2024-11-06 09:09:00.744835] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.822 [2024-11-06 09:09:00.744849] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.822 qpair failed and we were unable to recover it. 00:33:10.822 [2024-11-06 09:09:00.754764] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.822 [2024-11-06 09:09:00.754815] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.822 [2024-11-06 09:09:00.754829] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.822 [2024-11-06 09:09:00.754840] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.822 [2024-11-06 09:09:00.754846] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.822 [2024-11-06 09:09:00.754860] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.822 qpair failed and we were unable to recover it. 00:33:10.822 [2024-11-06 09:09:00.764750] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.764800] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.764814] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.764822] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.764828] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.764842] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.774808] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.774877] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.774890] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.774898] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.774904] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.774918] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.784744] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.784801] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.784817] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.784825] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.784831] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.784846] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.794865] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.794916] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.794930] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.794937] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.794944] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.794962] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.804891] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.804953] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.804967] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.804974] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.804981] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.804995] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.814930] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.814989] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.815003] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.815011] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.815017] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.815031] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.824847] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.824903] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.824919] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.824926] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.824933] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.824948] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.834989] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.835042] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.835057] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.835064] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.835070] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.835084] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.844991] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.845057] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.845071] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.845078] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.845085] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.845099] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.855048] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.855134] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.855147] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.855155] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.855162] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.855175] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.865091] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.865143] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.865157] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.865164] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.865170] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.865184] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.875081] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.875136] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.875150] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.875157] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.875164] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.875178] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.885127] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.823 [2024-11-06 09:09:00.885180] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.823 [2024-11-06 09:09:00.885194] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.823 [2024-11-06 09:09:00.885205] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.823 [2024-11-06 09:09:00.885212] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.823 [2024-11-06 09:09:00.885226] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.823 qpair failed and we were unable to recover it. 00:33:10.823 [2024-11-06 09:09:00.895148] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.824 [2024-11-06 09:09:00.895208] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.824 [2024-11-06 09:09:00.895222] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.824 [2024-11-06 09:09:00.895230] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.824 [2024-11-06 09:09:00.895236] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.824 [2024-11-06 09:09:00.895250] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.824 qpair failed and we were unable to recover it. 00:33:10.824 [2024-11-06 09:09:00.905155] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.824 [2024-11-06 09:09:00.905212] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.824 [2024-11-06 09:09:00.905225] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.824 [2024-11-06 09:09:00.905232] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.824 [2024-11-06 09:09:00.905239] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.824 [2024-11-06 09:09:00.905253] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.824 qpair failed and we were unable to recover it. 00:33:10.824 [2024-11-06 09:09:00.915162] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.824 [2024-11-06 09:09:00.915220] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.824 [2024-11-06 09:09:00.915233] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.824 [2024-11-06 09:09:00.915241] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.824 [2024-11-06 09:09:00.915247] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.824 [2024-11-06 09:09:00.915261] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.824 qpair failed and we were unable to recover it. 00:33:10.824 [2024-11-06 09:09:00.925239] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:10.824 [2024-11-06 09:09:00.925295] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:10.824 [2024-11-06 09:09:00.925308] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:10.824 [2024-11-06 09:09:00.925316] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:10.824 [2024-11-06 09:09:00.925322] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:10.824 [2024-11-06 09:09:00.925339] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:10.824 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:00.935158] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:00.935221] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:00.935234] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:00.935242] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:00.935248] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:00.935262] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:00.945302] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:00.945404] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:00.945417] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:00.945425] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:00.945432] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:00.945445] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:00.955211] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:00.955271] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:00.955284] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:00.955291] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:00.955298] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:00.955312] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:00.965359] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:00.965411] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:00.965424] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:00.965432] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:00.965439] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:00.965452] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:00.975374] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:00.975432] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:00.975446] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:00.975453] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:00.975460] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:00.975474] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:00.985386] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:00.985446] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:00.985460] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:00.985467] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:00.985474] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:00.985488] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:00.995449] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:00.995509] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:00.995523] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:00.995530] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:00.995537] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:00.995551] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:01.005453] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:01.005504] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:01.005518] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:01.005526] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:01.005532] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:01.005546] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:01.015491] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:01.015582] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:01.015596] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:01.015610] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:01.015616] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:01.015631] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:01.025579] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:01.025636] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:01.025650] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:01.025657] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:01.025664] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:01.025678] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:01.035551] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:01.035603] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:01.035617] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:01.035625] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:01.035632] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:01.035647] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:01.045568] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:01.045629] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:01.045643] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:01.045651] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:01.045657] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.087 [2024-11-06 09:09:01.045672] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.087 qpair failed and we were unable to recover it. 00:33:11.087 [2024-11-06 09:09:01.055593] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.087 [2024-11-06 09:09:01.055654] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.087 [2024-11-06 09:09:01.055668] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.087 [2024-11-06 09:09:01.055676] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.087 [2024-11-06 09:09:01.055682] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.055700] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.065641] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.065706] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.065720] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.065729] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.065736] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.065756] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.075613] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.075671] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.075684] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.075692] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.075698] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.075712] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.085672] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.085729] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.085742] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.085753] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.085760] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.085775] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.095706] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.095771] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.095785] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.095792] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.095799] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.095813] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.105752] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.105811] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.105825] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.105832] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.105839] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.105853] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.115743] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.115810] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.115825] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.115832] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.115839] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.115854] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.125780] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.125838] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.125851] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.125859] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.125865] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.125880] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.135865] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.135926] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.135940] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.135947] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.135954] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.135969] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.145847] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.145905] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.145919] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.145931] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.145937] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.145952] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.155846] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.155899] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.155912] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.155920] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.155926] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.155940] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.165921] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.165995] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.166009] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.166016] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.166022] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.166038] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.175931] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.175994] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.176008] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.176016] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.176022] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.088 [2024-11-06 09:09:01.176036] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.088 qpair failed and we were unable to recover it. 00:33:11.088 [2024-11-06 09:09:01.185985] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.088 [2024-11-06 09:09:01.186046] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.088 [2024-11-06 09:09:01.186059] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.088 [2024-11-06 09:09:01.186067] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.088 [2024-11-06 09:09:01.186073] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.089 [2024-11-06 09:09:01.186091] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.089 qpair failed and we were unable to recover it. 00:33:11.089 [2024-11-06 09:09:01.195989] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.089 [2024-11-06 09:09:01.196086] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.089 [2024-11-06 09:09:01.196101] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.089 [2024-11-06 09:09:01.196108] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.089 [2024-11-06 09:09:01.196114] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.089 [2024-11-06 09:09:01.196128] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.089 qpair failed and we were unable to recover it. 00:33:11.351 [2024-11-06 09:09:01.206008] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.351 [2024-11-06 09:09:01.206062] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.351 [2024-11-06 09:09:01.206076] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.351 [2024-11-06 09:09:01.206083] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.351 [2024-11-06 09:09:01.206090] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.351 [2024-11-06 09:09:01.206104] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.351 qpair failed and we were unable to recover it. 00:33:11.351 [2024-11-06 09:09:01.215916] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.351 [2024-11-06 09:09:01.215980] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.351 [2024-11-06 09:09:01.215994] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.351 [2024-11-06 09:09:01.216001] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.351 [2024-11-06 09:09:01.216008] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.351 [2024-11-06 09:09:01.216022] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.351 qpair failed and we were unable to recover it. 00:33:11.351 [2024-11-06 09:09:01.226047] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.351 [2024-11-06 09:09:01.226103] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.351 [2024-11-06 09:09:01.226117] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.351 [2024-11-06 09:09:01.226124] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.351 [2024-11-06 09:09:01.226131] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.351 [2024-11-06 09:09:01.226145] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.351 qpair failed and we were unable to recover it. 00:33:11.351 [2024-11-06 09:09:01.236096] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.351 [2024-11-06 09:09:01.236158] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.351 [2024-11-06 09:09:01.236173] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.351 [2024-11-06 09:09:01.236180] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.351 [2024-11-06 09:09:01.236187] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.351 [2024-11-06 09:09:01.236202] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.351 qpair failed and we were unable to recover it. 00:33:11.351 [2024-11-06 09:09:01.246123] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.351 [2024-11-06 09:09:01.246171] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.351 [2024-11-06 09:09:01.246185] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.351 [2024-11-06 09:09:01.246192] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.351 [2024-11-06 09:09:01.246198] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.351 [2024-11-06 09:09:01.246212] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.351 qpair failed and we were unable to recover it. 00:33:11.351 [2024-11-06 09:09:01.256150] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.351 [2024-11-06 09:09:01.256209] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.351 [2024-11-06 09:09:01.256223] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.351 [2024-11-06 09:09:01.256230] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.351 [2024-11-06 09:09:01.256237] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.351 [2024-11-06 09:09:01.256251] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.351 qpair failed and we were unable to recover it. 00:33:11.351 [2024-11-06 09:09:01.266241] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.351 [2024-11-06 09:09:01.266302] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.351 [2024-11-06 09:09:01.266316] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.351 [2024-11-06 09:09:01.266323] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.351 [2024-11-06 09:09:01.266330] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.351 [2024-11-06 09:09:01.266344] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.351 qpair failed and we were unable to recover it. 00:33:11.351 [2024-11-06 09:09:01.276214] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.351 [2024-11-06 09:09:01.276273] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.351 [2024-11-06 09:09:01.276287] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.351 [2024-11-06 09:09:01.276298] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.351 [2024-11-06 09:09:01.276304] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.351 [2024-11-06 09:09:01.276318] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.351 qpair failed and we were unable to recover it. 00:33:11.351 [2024-11-06 09:09:01.286232] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.351 [2024-11-06 09:09:01.286280] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.351 [2024-11-06 09:09:01.286294] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.351 [2024-11-06 09:09:01.286301] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.351 [2024-11-06 09:09:01.286308] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.351 [2024-11-06 09:09:01.286322] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.351 qpair failed and we were unable to recover it. 00:33:11.351 [2024-11-06 09:09:01.296277] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.351 [2024-11-06 09:09:01.296334] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.296347] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.296355] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.296361] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.296376] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.306303] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.306367] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.306381] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.306388] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.306395] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.306409] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.316297] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.316356] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.316370] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.316378] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.316384] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.316402] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.326338] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.326398] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.326412] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.326420] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.326426] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.326440] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.336360] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.336417] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.336431] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.336438] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.336445] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.336458] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.346418] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.346476] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.346490] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.346497] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.346504] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.346518] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.356428] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.356535] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.356561] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.356570] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.356577] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.356596] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.366448] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.366513] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.366540] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.366549] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.366557] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.366577] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.376365] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.376422] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.376438] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.376446] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.376452] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.376468] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.386528] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.386584] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.386598] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.386606] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.386613] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.386627] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.396532] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.396585] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.396599] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.396606] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.396613] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.396628] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.406567] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.406623] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.406637] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.406648] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.406655] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.406669] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.416598] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.416700] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.416714] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.352 [2024-11-06 09:09:01.416721] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.352 [2024-11-06 09:09:01.416729] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.352 [2024-11-06 09:09:01.416743] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.352 qpair failed and we were unable to recover it. 00:33:11.352 [2024-11-06 09:09:01.426614] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.352 [2024-11-06 09:09:01.426700] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.352 [2024-11-06 09:09:01.426714] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.353 [2024-11-06 09:09:01.426722] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.353 [2024-11-06 09:09:01.426729] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.353 [2024-11-06 09:09:01.426743] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.353 qpair failed and we were unable to recover it. 00:33:11.353 [2024-11-06 09:09:01.436640] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.353 [2024-11-06 09:09:01.436692] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.353 [2024-11-06 09:09:01.436706] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.353 [2024-11-06 09:09:01.436713] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.353 [2024-11-06 09:09:01.436720] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.353 [2024-11-06 09:09:01.436734] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.353 qpair failed and we were unable to recover it. 00:33:11.353 [2024-11-06 09:09:01.446636] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.353 [2024-11-06 09:09:01.446695] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.353 [2024-11-06 09:09:01.446710] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.353 [2024-11-06 09:09:01.446718] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.353 [2024-11-06 09:09:01.446724] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.353 [2024-11-06 09:09:01.446743] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.353 qpair failed and we were unable to recover it. 00:33:11.353 [2024-11-06 09:09:01.456698] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.353 [2024-11-06 09:09:01.456757] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.353 [2024-11-06 09:09:01.456771] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.353 [2024-11-06 09:09:01.456779] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.353 [2024-11-06 09:09:01.456786] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.353 [2024-11-06 09:09:01.456800] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.353 qpair failed and we were unable to recover it. 00:33:11.615 [2024-11-06 09:09:01.466743] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.615 [2024-11-06 09:09:01.466846] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.615 [2024-11-06 09:09:01.466860] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.615 [2024-11-06 09:09:01.466868] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.615 [2024-11-06 09:09:01.466874] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.615 [2024-11-06 09:09:01.466889] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.615 qpair failed and we were unable to recover it. 00:33:11.615 [2024-11-06 09:09:01.476759] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.615 [2024-11-06 09:09:01.476846] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.615 [2024-11-06 09:09:01.476860] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.615 [2024-11-06 09:09:01.476868] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.615 [2024-11-06 09:09:01.476875] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.615 [2024-11-06 09:09:01.476889] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.615 qpair failed and we were unable to recover it. 00:33:11.615 [2024-11-06 09:09:01.486707] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.615 [2024-11-06 09:09:01.486803] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.615 [2024-11-06 09:09:01.486818] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.615 [2024-11-06 09:09:01.486825] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.615 [2024-11-06 09:09:01.486832] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.615 [2024-11-06 09:09:01.486846] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.615 qpair failed and we were unable to recover it. 00:33:11.615 [2024-11-06 09:09:01.496820] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.615 [2024-11-06 09:09:01.496880] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.615 [2024-11-06 09:09:01.496894] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.615 [2024-11-06 09:09:01.496901] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.615 [2024-11-06 09:09:01.496908] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.615 [2024-11-06 09:09:01.496922] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.615 qpair failed and we were unable to recover it. 00:33:11.615 [2024-11-06 09:09:01.506811] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.615 [2024-11-06 09:09:01.506870] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.615 [2024-11-06 09:09:01.506884] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.615 [2024-11-06 09:09:01.506891] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.615 [2024-11-06 09:09:01.506898] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.615 [2024-11-06 09:09:01.506912] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.615 qpair failed and we were unable to recover it. 00:33:11.615 [2024-11-06 09:09:01.516867] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.615 [2024-11-06 09:09:01.516919] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.615 [2024-11-06 09:09:01.516932] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.615 [2024-11-06 09:09:01.516940] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.615 [2024-11-06 09:09:01.516946] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.615 [2024-11-06 09:09:01.516961] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.615 qpair failed and we were unable to recover it. 00:33:11.615 [2024-11-06 09:09:01.526829] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.615 [2024-11-06 09:09:01.526877] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.615 [2024-11-06 09:09:01.526891] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.615 [2024-11-06 09:09:01.526898] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.615 [2024-11-06 09:09:01.526905] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.615 [2024-11-06 09:09:01.526920] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.615 qpair failed and we were unable to recover it. 00:33:11.615 [2024-11-06 09:09:01.536908] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.536968] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.536982] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.536997] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.537004] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.537019] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.546979] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.547068] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.547082] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.547090] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.547097] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.547111] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.556985] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.557045] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.557059] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.557066] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.557072] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.557086] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.566986] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.567032] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.567045] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.567052] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.567059] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.567073] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.577030] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.577085] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.577099] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.577106] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.577113] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.577130] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.587052] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.587112] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.587126] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.587133] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.587140] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.587154] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.597102] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.597160] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.597174] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.597181] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.597188] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.597202] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.607085] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.607137] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.607151] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.607158] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.607165] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.607180] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.617125] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.617183] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.617197] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.617205] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.617211] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.617225] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.627197] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.627300] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.627314] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.627321] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.627327] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.627342] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.637205] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.637255] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.637269] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.637276] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.637283] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.637296] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.647191] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.647257] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.647271] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.647278] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.647285] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.647298] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.657236] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.657295] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.616 [2024-11-06 09:09:01.657309] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.616 [2024-11-06 09:09:01.657317] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.616 [2024-11-06 09:09:01.657323] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.616 [2024-11-06 09:09:01.657338] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.616 qpair failed and we were unable to recover it. 00:33:11.616 [2024-11-06 09:09:01.667313] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.616 [2024-11-06 09:09:01.667372] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.617 [2024-11-06 09:09:01.667386] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.617 [2024-11-06 09:09:01.667396] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.617 [2024-11-06 09:09:01.667403] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.617 [2024-11-06 09:09:01.667417] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.617 qpair failed and we were unable to recover it. 00:33:11.617 [2024-11-06 09:09:01.677318] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.617 [2024-11-06 09:09:01.677379] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.617 [2024-11-06 09:09:01.677393] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.617 [2024-11-06 09:09:01.677401] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.617 [2024-11-06 09:09:01.677407] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.617 [2024-11-06 09:09:01.677421] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.617 qpair failed and we were unable to recover it. 00:33:11.617 [2024-11-06 09:09:01.687184] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.617 [2024-11-06 09:09:01.687234] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.617 [2024-11-06 09:09:01.687248] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.617 [2024-11-06 09:09:01.687256] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.617 [2024-11-06 09:09:01.687263] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.617 [2024-11-06 09:09:01.687277] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.617 qpair failed and we were unable to recover it. 00:33:11.617 [2024-11-06 09:09:01.697417] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.617 [2024-11-06 09:09:01.697476] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.617 [2024-11-06 09:09:01.697489] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.617 [2024-11-06 09:09:01.697497] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.617 [2024-11-06 09:09:01.697504] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.617 [2024-11-06 09:09:01.697517] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.617 qpair failed and we were unable to recover it. 00:33:11.617 [2024-11-06 09:09:01.707419] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.617 [2024-11-06 09:09:01.707473] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.617 [2024-11-06 09:09:01.707487] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.617 [2024-11-06 09:09:01.707495] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.617 [2024-11-06 09:09:01.707501] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.617 [2024-11-06 09:09:01.707519] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.617 qpair failed and we were unable to recover it. 00:33:11.617 [2024-11-06 09:09:01.717365] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.617 [2024-11-06 09:09:01.717427] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.617 [2024-11-06 09:09:01.717454] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.617 [2024-11-06 09:09:01.717462] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.617 [2024-11-06 09:09:01.717470] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.617 [2024-11-06 09:09:01.717490] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.617 qpair failed and we were unable to recover it. 00:33:11.879 [2024-11-06 09:09:01.727408] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.879 [2024-11-06 09:09:01.727455] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.879 [2024-11-06 09:09:01.727471] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.879 [2024-11-06 09:09:01.727478] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.879 [2024-11-06 09:09:01.727485] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.879 [2024-11-06 09:09:01.727500] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.879 qpair failed and we were unable to recover it. 00:33:11.879 [2024-11-06 09:09:01.737490] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.879 [2024-11-06 09:09:01.737554] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.879 [2024-11-06 09:09:01.737580] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.879 [2024-11-06 09:09:01.737589] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.879 [2024-11-06 09:09:01.737597] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.879 [2024-11-06 09:09:01.737617] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.879 qpair failed and we were unable to recover it. 00:33:11.879 [2024-11-06 09:09:01.747516] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.879 [2024-11-06 09:09:01.747574] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.879 [2024-11-06 09:09:01.747589] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.879 [2024-11-06 09:09:01.747597] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.879 [2024-11-06 09:09:01.747604] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.879 [2024-11-06 09:09:01.747619] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.879 qpair failed and we were unable to recover it. 00:33:11.879 [2024-11-06 09:09:01.757487] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.879 [2024-11-06 09:09:01.757547] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.879 [2024-11-06 09:09:01.757561] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.879 [2024-11-06 09:09:01.757568] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.879 [2024-11-06 09:09:01.757575] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.879 [2024-11-06 09:09:01.757589] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.879 qpair failed and we were unable to recover it. 00:33:11.879 [2024-11-06 09:09:01.767397] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.879 [2024-11-06 09:09:01.767454] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.879 [2024-11-06 09:09:01.767467] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.879 [2024-11-06 09:09:01.767475] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.879 [2024-11-06 09:09:01.767481] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.879 [2024-11-06 09:09:01.767496] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.879 qpair failed and we were unable to recover it. 00:33:11.879 [2024-11-06 09:09:01.777556] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.879 [2024-11-06 09:09:01.777618] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.879 [2024-11-06 09:09:01.777633] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.879 [2024-11-06 09:09:01.777640] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.879 [2024-11-06 09:09:01.777647] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.879 [2024-11-06 09:09:01.777662] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.879 qpair failed and we were unable to recover it. 00:33:11.879 [2024-11-06 09:09:01.787614] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.879 [2024-11-06 09:09:01.787670] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.879 [2024-11-06 09:09:01.787684] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.879 [2024-11-06 09:09:01.787691] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.879 [2024-11-06 09:09:01.787698] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.879 [2024-11-06 09:09:01.787712] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.879 qpair failed and we were unable to recover it. 00:33:11.879 [2024-11-06 09:09:01.797601] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.879 [2024-11-06 09:09:01.797646] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.879 [2024-11-06 09:09:01.797660] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.879 [2024-11-06 09:09:01.797672] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.879 [2024-11-06 09:09:01.797679] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.879 [2024-11-06 09:09:01.797693] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.879 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.807619] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.807668] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.807682] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.807690] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.807696] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.807711] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.817692] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.817750] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.817764] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.817772] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.817778] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.817792] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.827735] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.827798] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.827812] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.827819] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.827826] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.827840] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.837696] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.837752] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.837765] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.837773] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.837779] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.837797] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.847731] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.847780] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.847794] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.847802] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.847808] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.847823] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.857822] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.857881] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.857894] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.857902] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.857908] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.857922] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.867830] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.867885] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.867900] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.867907] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.867914] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.867929] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.877816] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.877869] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.877883] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.877891] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.877897] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.877911] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.887833] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.887891] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.887905] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.887914] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.887921] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.887936] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.897954] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.898016] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.898030] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.898038] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.898045] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.898059] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.907939] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.908001] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.908016] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.908024] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.908030] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.908044] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.917927] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.918021] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.918035] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.918043] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.918049] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.918063] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.927977] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.880 [2024-11-06 09:09:01.928066] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.880 [2024-11-06 09:09:01.928079] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.880 [2024-11-06 09:09:01.928092] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.880 [2024-11-06 09:09:01.928098] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.880 [2024-11-06 09:09:01.928112] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.880 qpair failed and we were unable to recover it. 00:33:11.880 [2024-11-06 09:09:01.938035] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.881 [2024-11-06 09:09:01.938124] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.881 [2024-11-06 09:09:01.938139] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.881 [2024-11-06 09:09:01.938147] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.881 [2024-11-06 09:09:01.938154] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.881 [2024-11-06 09:09:01.938171] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.881 qpair failed and we were unable to recover it. 00:33:11.881 [2024-11-06 09:09:01.947984] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.881 [2024-11-06 09:09:01.948042] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.881 [2024-11-06 09:09:01.948056] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.881 [2024-11-06 09:09:01.948064] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.881 [2024-11-06 09:09:01.948071] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.881 [2024-11-06 09:09:01.948085] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.881 qpair failed and we were unable to recover it. 00:33:11.881 [2024-11-06 09:09:01.958010] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.881 [2024-11-06 09:09:01.958065] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.881 [2024-11-06 09:09:01.958078] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.881 [2024-11-06 09:09:01.958086] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.881 [2024-11-06 09:09:01.958092] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.881 [2024-11-06 09:09:01.958106] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.881 qpair failed and we were unable to recover it. 00:33:11.881 [2024-11-06 09:09:01.968030] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.881 [2024-11-06 09:09:01.968081] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.881 [2024-11-06 09:09:01.968095] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.881 [2024-11-06 09:09:01.968102] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.881 [2024-11-06 09:09:01.968109] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.881 [2024-11-06 09:09:01.968127] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.881 qpair failed and we were unable to recover it. 00:33:11.881 [2024-11-06 09:09:01.978133] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.881 [2024-11-06 09:09:01.978188] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.881 [2024-11-06 09:09:01.978202] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.881 [2024-11-06 09:09:01.978210] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.881 [2024-11-06 09:09:01.978216] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.881 [2024-11-06 09:09:01.978230] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.881 qpair failed and we were unable to recover it. 00:33:11.881 [2024-11-06 09:09:01.988172] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:11.881 [2024-11-06 09:09:01.988226] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:11.881 [2024-11-06 09:09:01.988239] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:11.881 [2024-11-06 09:09:01.988247] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:11.881 [2024-11-06 09:09:01.988253] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:11.881 [2024-11-06 09:09:01.988267] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:11.881 qpair failed and we were unable to recover it. 00:33:12.143 [2024-11-06 09:09:01.998270] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.143 [2024-11-06 09:09:01.998324] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.143 [2024-11-06 09:09:01.998338] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.143 [2024-11-06 09:09:01.998346] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.143 [2024-11-06 09:09:01.998352] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.143 [2024-11-06 09:09:01.998366] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.143 qpair failed and we were unable to recover it. 00:33:12.143 [2024-11-06 09:09:02.008096] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.143 [2024-11-06 09:09:02.008150] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.143 [2024-11-06 09:09:02.008165] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.143 [2024-11-06 09:09:02.008173] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.143 [2024-11-06 09:09:02.008180] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.143 [2024-11-06 09:09:02.008194] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.143 qpair failed and we were unable to recover it. 00:33:12.143 [2024-11-06 09:09:02.018275] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.143 [2024-11-06 09:09:02.018334] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.143 [2024-11-06 09:09:02.018348] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.143 [2024-11-06 09:09:02.018355] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.143 [2024-11-06 09:09:02.018362] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.143 [2024-11-06 09:09:02.018376] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.143 qpair failed and we were unable to recover it. 00:33:12.143 [2024-11-06 09:09:02.028315] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.143 [2024-11-06 09:09:02.028374] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.143 [2024-11-06 09:09:02.028387] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.143 [2024-11-06 09:09:02.028394] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.143 [2024-11-06 09:09:02.028401] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.143 [2024-11-06 09:09:02.028414] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.143 qpair failed and we were unable to recover it. 00:33:12.143 [2024-11-06 09:09:02.038234] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.143 [2024-11-06 09:09:02.038283] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.143 [2024-11-06 09:09:02.038296] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.143 [2024-11-06 09:09:02.038304] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.143 [2024-11-06 09:09:02.038310] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.143 [2024-11-06 09:09:02.038324] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.143 qpair failed and we were unable to recover it. 00:33:12.143 [2024-11-06 09:09:02.048329] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.143 [2024-11-06 09:09:02.048377] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.143 [2024-11-06 09:09:02.048390] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.143 [2024-11-06 09:09:02.048398] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.143 [2024-11-06 09:09:02.048404] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.143 [2024-11-06 09:09:02.048418] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.143 qpair failed and we were unable to recover it. 00:33:12.143 [2024-11-06 09:09:02.058264] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.143 [2024-11-06 09:09:02.058356] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.143 [2024-11-06 09:09:02.058370] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.143 [2024-11-06 09:09:02.058382] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.143 [2024-11-06 09:09:02.058388] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.058403] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.068381] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.068444] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.068459] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.068467] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.068473] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.068488] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.078355] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.078411] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.078427] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.078436] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.078442] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.078460] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.088405] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.088451] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.088466] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.088474] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.088481] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.088495] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.098429] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.098488] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.098503] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.098511] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.098518] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.098536] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.108498] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.108556] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.108570] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.108577] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.108584] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.108598] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.118470] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.118527] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.118553] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.118563] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.118570] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.118589] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.128495] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.128550] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.128576] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.128588] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.128595] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.128615] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.138556] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.138614] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.138630] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.138638] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.138644] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.138660] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.148610] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.148667] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.148682] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.148689] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.148696] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.148710] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.158583] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.158637] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.158651] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.158658] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.158665] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.158679] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.168607] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.168689] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.168703] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.168711] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.168718] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.168732] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.178682] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.178741] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.144 [2024-11-06 09:09:02.178759] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.144 [2024-11-06 09:09:02.178767] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.144 [2024-11-06 09:09:02.178773] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.144 [2024-11-06 09:09:02.178788] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.144 qpair failed and we were unable to recover it. 00:33:12.144 [2024-11-06 09:09:02.188718] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.144 [2024-11-06 09:09:02.188781] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.145 [2024-11-06 09:09:02.188794] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.145 [2024-11-06 09:09:02.188806] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.145 [2024-11-06 09:09:02.188812] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.145 [2024-11-06 09:09:02.188827] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.145 qpair failed and we were unable to recover it. 00:33:12.145 [2024-11-06 09:09:02.198693] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.145 [2024-11-06 09:09:02.198791] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.145 [2024-11-06 09:09:02.198805] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.145 [2024-11-06 09:09:02.198813] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.145 [2024-11-06 09:09:02.198820] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.145 [2024-11-06 09:09:02.198834] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.145 qpair failed and we were unable to recover it. 00:33:12.145 [2024-11-06 09:09:02.208694] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.145 [2024-11-06 09:09:02.208791] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.145 [2024-11-06 09:09:02.208805] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.145 [2024-11-06 09:09:02.208812] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.145 [2024-11-06 09:09:02.208818] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.145 [2024-11-06 09:09:02.208832] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.145 qpair failed and we were unable to recover it. 00:33:12.145 [2024-11-06 09:09:02.218787] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.145 [2024-11-06 09:09:02.218845] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.145 [2024-11-06 09:09:02.218858] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.145 [2024-11-06 09:09:02.218866] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.145 [2024-11-06 09:09:02.218872] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.145 [2024-11-06 09:09:02.218886] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.145 qpair failed and we were unable to recover it. 00:33:12.145 [2024-11-06 09:09:02.228820] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.145 [2024-11-06 09:09:02.228882] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.145 [2024-11-06 09:09:02.228895] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.145 [2024-11-06 09:09:02.228903] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.145 [2024-11-06 09:09:02.228909] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.145 [2024-11-06 09:09:02.228927] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.145 qpair failed and we were unable to recover it. 00:33:12.145 [2024-11-06 09:09:02.238804] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.145 [2024-11-06 09:09:02.238851] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.145 [2024-11-06 09:09:02.238864] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.145 [2024-11-06 09:09:02.238872] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.145 [2024-11-06 09:09:02.238878] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.145 [2024-11-06 09:09:02.238892] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.145 qpair failed and we were unable to recover it. 00:33:12.145 [2024-11-06 09:09:02.248835] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.145 [2024-11-06 09:09:02.248887] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.145 [2024-11-06 09:09:02.248901] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.145 [2024-11-06 09:09:02.248908] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.145 [2024-11-06 09:09:02.248915] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.145 [2024-11-06 09:09:02.248929] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.145 qpair failed and we were unable to recover it. 00:33:12.407 [2024-11-06 09:09:02.258895] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.407 [2024-11-06 09:09:02.258950] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.407 [2024-11-06 09:09:02.258964] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.407 [2024-11-06 09:09:02.258971] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.407 [2024-11-06 09:09:02.258978] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.407 [2024-11-06 09:09:02.258992] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.407 qpair failed and we were unable to recover it. 00:33:12.407 [2024-11-06 09:09:02.268934] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.407 [2024-11-06 09:09:02.268990] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.407 [2024-11-06 09:09:02.269003] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.407 [2024-11-06 09:09:02.269011] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.407 [2024-11-06 09:09:02.269017] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.407 [2024-11-06 09:09:02.269031] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.407 qpair failed and we were unable to recover it. 00:33:12.407 [2024-11-06 09:09:02.278889] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.407 [2024-11-06 09:09:02.278940] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.407 [2024-11-06 09:09:02.278956] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.407 [2024-11-06 09:09:02.278963] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.407 [2024-11-06 09:09:02.278969] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.407 [2024-11-06 09:09:02.278984] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.407 qpair failed and we were unable to recover it. 00:33:12.407 [2024-11-06 09:09:02.288956] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.407 [2024-11-06 09:09:02.289008] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.407 [2024-11-06 09:09:02.289021] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.407 [2024-11-06 09:09:02.289029] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.407 [2024-11-06 09:09:02.289035] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.407 [2024-11-06 09:09:02.289049] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.407 qpair failed and we were unable to recover it. 00:33:12.407 [2024-11-06 09:09:02.299007] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.407 [2024-11-06 09:09:02.299064] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.407 [2024-11-06 09:09:02.299078] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.407 [2024-11-06 09:09:02.299085] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.299092] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.299106] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.308927] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.308982] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.308995] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.309003] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.309009] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.309023] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.319031] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.319081] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.319095] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.319106] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.319112] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.319126] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.328929] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.328978] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.328991] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.328999] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.329005] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.329019] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.339239] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.339317] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.339331] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.339338] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.339345] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.339360] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.349153] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.349206] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.349219] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.349226] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.349233] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.349246] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.359124] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.359174] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.359187] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.359194] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.359201] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.359218] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.369155] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.369241] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.369255] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.369264] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.369271] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.369285] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.379263] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.379322] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.379336] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.379343] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.379350] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.379364] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.389252] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.389313] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.389327] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.389334] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.389341] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.389355] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.399234] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.399285] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.399298] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.399305] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.399312] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.399326] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.409257] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.409311] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.409325] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.409332] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.409339] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.409353] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.419341] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.419395] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.419409] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.419417] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.419423] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.408 [2024-11-06 09:09:02.419438] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.408 qpair failed and we were unable to recover it. 00:33:12.408 [2024-11-06 09:09:02.429383] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.408 [2024-11-06 09:09:02.429437] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.408 [2024-11-06 09:09:02.429451] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.408 [2024-11-06 09:09:02.429458] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.408 [2024-11-06 09:09:02.429465] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.409 [2024-11-06 09:09:02.429478] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.409 qpair failed and we were unable to recover it. 00:33:12.409 [2024-11-06 09:09:02.439315] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.409 [2024-11-06 09:09:02.439385] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.409 [2024-11-06 09:09:02.439400] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.409 [2024-11-06 09:09:02.439407] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.409 [2024-11-06 09:09:02.439414] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.409 [2024-11-06 09:09:02.439428] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.409 qpair failed and we were unable to recover it. 00:33:12.409 [2024-11-06 09:09:02.449361] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.409 [2024-11-06 09:09:02.449413] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.409 [2024-11-06 09:09:02.449427] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.409 [2024-11-06 09:09:02.449438] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.409 [2024-11-06 09:09:02.449445] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.409 [2024-11-06 09:09:02.449459] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.409 qpair failed and we were unable to recover it. 00:33:12.409 [2024-11-06 09:09:02.459491] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.409 [2024-11-06 09:09:02.459567] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.409 [2024-11-06 09:09:02.459581] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.409 [2024-11-06 09:09:02.459588] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.409 [2024-11-06 09:09:02.459595] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.409 [2024-11-06 09:09:02.459609] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.409 qpair failed and we were unable to recover it. 00:33:12.409 [2024-11-06 09:09:02.469481] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.409 [2024-11-06 09:09:02.469535] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.409 [2024-11-06 09:09:02.469548] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.409 [2024-11-06 09:09:02.469556] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.409 [2024-11-06 09:09:02.469562] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.409 [2024-11-06 09:09:02.469576] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.409 qpair failed and we were unable to recover it. 00:33:12.409 [2024-11-06 09:09:02.479464] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.409 [2024-11-06 09:09:02.479512] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.409 [2024-11-06 09:09:02.479526] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.409 [2024-11-06 09:09:02.479533] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.409 [2024-11-06 09:09:02.479539] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.409 [2024-11-06 09:09:02.479553] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.409 qpair failed and we were unable to recover it. 00:33:12.409 [2024-11-06 09:09:02.489483] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.409 [2024-11-06 09:09:02.489536] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.409 [2024-11-06 09:09:02.489550] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.409 [2024-11-06 09:09:02.489558] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.409 [2024-11-06 09:09:02.489565] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.409 [2024-11-06 09:09:02.489586] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.409 qpair failed and we were unable to recover it. 00:33:12.409 [2024-11-06 09:09:02.499586] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.409 [2024-11-06 09:09:02.499684] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.409 [2024-11-06 09:09:02.499698] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.409 [2024-11-06 09:09:02.499706] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.409 [2024-11-06 09:09:02.499713] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.409 [2024-11-06 09:09:02.499727] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.409 qpair failed and we were unable to recover it. 00:33:12.409 [2024-11-06 09:09:02.509639] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.409 [2024-11-06 09:09:02.509725] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.409 [2024-11-06 09:09:02.509739] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.409 [2024-11-06 09:09:02.509750] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.409 [2024-11-06 09:09:02.509757] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.409 [2024-11-06 09:09:02.509771] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.409 qpair failed and we were unable to recover it. 00:33:12.671 [2024-11-06 09:09:02.519582] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.671 [2024-11-06 09:09:02.519631] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.671 [2024-11-06 09:09:02.519645] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.671 [2024-11-06 09:09:02.519652] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.671 [2024-11-06 09:09:02.519660] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.671 [2024-11-06 09:09:02.519674] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.671 qpair failed and we were unable to recover it. 00:33:12.671 [2024-11-06 09:09:02.529476] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.671 [2024-11-06 09:09:02.529570] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.671 [2024-11-06 09:09:02.529586] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.671 [2024-11-06 09:09:02.529594] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.671 [2024-11-06 09:09:02.529600] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.671 [2024-11-06 09:09:02.529615] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.671 qpair failed and we were unable to recover it. 00:33:12.671 [2024-11-06 09:09:02.539671] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.671 [2024-11-06 09:09:02.539730] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.671 [2024-11-06 09:09:02.539744] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.671 [2024-11-06 09:09:02.539755] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.671 [2024-11-06 09:09:02.539761] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.671 [2024-11-06 09:09:02.539776] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.671 qpair failed and we were unable to recover it. 00:33:12.671 [2024-11-06 09:09:02.549687] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.671 [2024-11-06 09:09:02.549750] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.671 [2024-11-06 09:09:02.549763] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.671 [2024-11-06 09:09:02.549770] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.671 [2024-11-06 09:09:02.549777] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.671 [2024-11-06 09:09:02.549791] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.671 qpair failed and we were unable to recover it. 00:33:12.671 [2024-11-06 09:09:02.559667] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.671 [2024-11-06 09:09:02.559717] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.671 [2024-11-06 09:09:02.559730] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.671 [2024-11-06 09:09:02.559738] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.671 [2024-11-06 09:09:02.559744] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.671 [2024-11-06 09:09:02.559763] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.671 qpair failed and we were unable to recover it. 00:33:12.671 [2024-11-06 09:09:02.569684] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.671 [2024-11-06 09:09:02.569737] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.671 [2024-11-06 09:09:02.569754] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.671 [2024-11-06 09:09:02.569762] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.671 [2024-11-06 09:09:02.569768] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.671 [2024-11-06 09:09:02.569782] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.671 qpair failed and we were unable to recover it. 00:33:12.671 [2024-11-06 09:09:02.579658] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.671 [2024-11-06 09:09:02.579718] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.671 [2024-11-06 09:09:02.579732] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.671 [2024-11-06 09:09:02.579744] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.671 [2024-11-06 09:09:02.579754] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.671 [2024-11-06 09:09:02.579769] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.671 qpair failed and we were unable to recover it. 00:33:12.671 [2024-11-06 09:09:02.589810] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.671 [2024-11-06 09:09:02.589866] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.671 [2024-11-06 09:09:02.589881] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.671 [2024-11-06 09:09:02.589888] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.671 [2024-11-06 09:09:02.589894] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.671 [2024-11-06 09:09:02.589908] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.671 qpair failed and we were unable to recover it. 00:33:12.671 [2024-11-06 09:09:02.599787] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.671 [2024-11-06 09:09:02.599835] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.671 [2024-11-06 09:09:02.599849] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.671 [2024-11-06 09:09:02.599856] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.671 [2024-11-06 09:09:02.599862] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.671 [2024-11-06 09:09:02.599876] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.671 qpair failed and we were unable to recover it. 00:33:12.671 [2024-11-06 09:09:02.609816] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.609868] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.609882] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.609889] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.609895] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.609910] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.619838] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.619897] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.619910] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.619918] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.619925] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.619942] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.629992] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.630093] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.630107] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.630114] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.630121] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.630135] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.639916] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.640004] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.640017] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.640025] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.640032] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.640045] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.649943] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.650030] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.650043] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.650051] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.650058] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.650071] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.660019] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.660075] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.660088] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.660096] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.660102] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.660116] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.670060] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.670114] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.670128] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.670135] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.670142] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.670156] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.680037] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.680092] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.680105] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.680113] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.680119] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.680133] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.690067] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.690112] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.690125] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.690132] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.690139] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.690153] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.700128] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.700188] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.700203] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.700210] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.700217] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.700231] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.710159] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.710215] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.710229] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.710240] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.710246] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.710260] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.720122] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.720168] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.720182] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.720189] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.720195] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.720209] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.730160] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.730210] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.672 [2024-11-06 09:09:02.730224] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.672 [2024-11-06 09:09:02.730231] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.672 [2024-11-06 09:09:02.730237] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.672 [2024-11-06 09:09:02.730251] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.672 qpair failed and we were unable to recover it. 00:33:12.672 [2024-11-06 09:09:02.740240] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.672 [2024-11-06 09:09:02.740324] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.673 [2024-11-06 09:09:02.740339] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.673 [2024-11-06 09:09:02.740347] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.673 [2024-11-06 09:09:02.740354] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.673 [2024-11-06 09:09:02.740370] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.673 qpair failed and we were unable to recover it. 00:33:12.673 [2024-11-06 09:09:02.750244] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.673 [2024-11-06 09:09:02.750299] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.673 [2024-11-06 09:09:02.750313] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.673 [2024-11-06 09:09:02.750320] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.673 [2024-11-06 09:09:02.750327] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.673 [2024-11-06 09:09:02.750344] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.673 qpair failed and we were unable to recover it. 00:33:12.673 [2024-11-06 09:09:02.760234] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.673 [2024-11-06 09:09:02.760284] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.673 [2024-11-06 09:09:02.760298] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.673 [2024-11-06 09:09:02.760306] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.673 [2024-11-06 09:09:02.760312] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.673 [2024-11-06 09:09:02.760326] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.673 qpair failed and we were unable to recover it. 00:33:12.673 [2024-11-06 09:09:02.770231] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.673 [2024-11-06 09:09:02.770277] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.673 [2024-11-06 09:09:02.770291] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.673 [2024-11-06 09:09:02.770298] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.673 [2024-11-06 09:09:02.770305] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.673 [2024-11-06 09:09:02.770319] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.673 qpair failed and we were unable to recover it. 00:33:12.673 [2024-11-06 09:09:02.780358] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.673 [2024-11-06 09:09:02.780411] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.673 [2024-11-06 09:09:02.780425] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.673 [2024-11-06 09:09:02.780432] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.673 [2024-11-06 09:09:02.780439] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.673 [2024-11-06 09:09:02.780453] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.673 qpair failed and we were unable to recover it. 00:33:12.934 [2024-11-06 09:09:02.790419] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.934 [2024-11-06 09:09:02.790501] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.934 [2024-11-06 09:09:02.790515] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.934 [2024-11-06 09:09:02.790522] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.934 [2024-11-06 09:09:02.790530] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.934 [2024-11-06 09:09:02.790543] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.934 qpair failed and we were unable to recover it. 00:33:12.934 [2024-11-06 09:09:02.800357] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.934 [2024-11-06 09:09:02.800414] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.934 [2024-11-06 09:09:02.800439] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.934 [2024-11-06 09:09:02.800448] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.934 [2024-11-06 09:09:02.800455] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.934 [2024-11-06 09:09:02.800475] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.934 qpair failed and we were unable to recover it. 00:33:12.934 [2024-11-06 09:09:02.810389] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.934 [2024-11-06 09:09:02.810442] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.934 [2024-11-06 09:09:02.810458] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.934 [2024-11-06 09:09:02.810466] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.934 [2024-11-06 09:09:02.810473] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.934 [2024-11-06 09:09:02.810488] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.934 qpair failed and we were unable to recover it. 00:33:12.934 [2024-11-06 09:09:02.820464] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.934 [2024-11-06 09:09:02.820573] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.934 [2024-11-06 09:09:02.820587] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.934 [2024-11-06 09:09:02.820595] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.934 [2024-11-06 09:09:02.820602] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.934 [2024-11-06 09:09:02.820616] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.934 qpair failed and we were unable to recover it. 00:33:12.934 [2024-11-06 09:09:02.830496] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.830552] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.830566] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.830573] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.830580] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.830594] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.840458] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.840510] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.840524] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.840535] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.840542] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.840557] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.850489] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.850542] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.850556] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.850563] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.850570] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.850585] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.860565] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.860621] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.860635] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.860642] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.860648] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.860662] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.870482] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.870544] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.870558] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.870566] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.870572] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.870586] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.880562] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.880613] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.880628] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.880635] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.880641] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.880659] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.890590] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.890642] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.890656] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.890664] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.890671] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.890685] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.900667] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.900723] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.900739] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.900750] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.900758] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.900773] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.910704] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.910760] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.910774] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.910782] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.910788] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.910803] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.920700] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.920753] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.920767] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.920775] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.920781] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.920795] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.930724] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.930777] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.930791] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.930798] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.930805] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.930819] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.940783] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.940838] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.940852] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.940860] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.940867] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.940881] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.950808] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.950864] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.950877] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.935 [2024-11-06 09:09:02.950885] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.935 [2024-11-06 09:09:02.950891] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.935 [2024-11-06 09:09:02.950905] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.935 qpair failed and we were unable to recover it. 00:33:12.935 [2024-11-06 09:09:02.960849] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.935 [2024-11-06 09:09:02.960903] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.935 [2024-11-06 09:09:02.960917] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.936 [2024-11-06 09:09:02.960924] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.936 [2024-11-06 09:09:02.960931] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.936 [2024-11-06 09:09:02.960945] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.936 qpair failed and we were unable to recover it. 00:33:12.936 [2024-11-06 09:09:02.970827] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.936 [2024-11-06 09:09:02.970907] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.936 [2024-11-06 09:09:02.970920] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.936 [2024-11-06 09:09:02.970931] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.936 [2024-11-06 09:09:02.970938] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.936 [2024-11-06 09:09:02.970952] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.936 qpair failed and we were unable to recover it. 00:33:12.936 [2024-11-06 09:09:02.980852] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.936 [2024-11-06 09:09:02.980911] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.936 [2024-11-06 09:09:02.980926] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.936 [2024-11-06 09:09:02.980933] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.936 [2024-11-06 09:09:02.980940] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.936 [2024-11-06 09:09:02.980953] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.936 qpair failed and we were unable to recover it. 00:33:12.936 [2024-11-06 09:09:02.990942] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.936 [2024-11-06 09:09:02.990997] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.936 [2024-11-06 09:09:02.991010] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.936 [2024-11-06 09:09:02.991018] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.936 [2024-11-06 09:09:02.991024] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.936 [2024-11-06 09:09:02.991038] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.936 qpair failed and we were unable to recover it. 00:33:12.936 [2024-11-06 09:09:03.000919] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.936 [2024-11-06 09:09:03.000966] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.936 [2024-11-06 09:09:03.000980] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.936 [2024-11-06 09:09:03.000988] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.936 [2024-11-06 09:09:03.000994] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.936 [2024-11-06 09:09:03.001008] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.936 qpair failed and we were unable to recover it. 00:33:12.936 [2024-11-06 09:09:03.010927] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.936 [2024-11-06 09:09:03.010977] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.936 [2024-11-06 09:09:03.010990] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.936 [2024-11-06 09:09:03.010998] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.936 [2024-11-06 09:09:03.011004] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.936 [2024-11-06 09:09:03.011022] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.936 qpair failed and we were unable to recover it. 00:33:12.936 [2024-11-06 09:09:03.021052] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.936 [2024-11-06 09:09:03.021111] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.936 [2024-11-06 09:09:03.021125] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.936 [2024-11-06 09:09:03.021132] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.936 [2024-11-06 09:09:03.021139] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.936 [2024-11-06 09:09:03.021153] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.936 qpair failed and we were unable to recover it. 00:33:12.936 [2024-11-06 09:09:03.031032] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.936 [2024-11-06 09:09:03.031082] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.936 [2024-11-06 09:09:03.031096] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.936 [2024-11-06 09:09:03.031103] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.936 [2024-11-06 09:09:03.031110] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.936 [2024-11-06 09:09:03.031124] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.936 qpair failed and we were unable to recover it. 00:33:12.936 [2024-11-06 09:09:03.041019] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:12.936 [2024-11-06 09:09:03.041076] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:12.936 [2024-11-06 09:09:03.041089] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:12.936 [2024-11-06 09:09:03.041097] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:12.936 [2024-11-06 09:09:03.041103] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:12.936 [2024-11-06 09:09:03.041117] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:12.936 qpair failed and we were unable to recover it. 00:33:13.197 [2024-11-06 09:09:03.051034] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.197 [2024-11-06 09:09:03.051083] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.197 [2024-11-06 09:09:03.051096] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.197 [2024-11-06 09:09:03.051104] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.197 [2024-11-06 09:09:03.051110] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.197 [2024-11-06 09:09:03.051124] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.197 qpair failed and we were unable to recover it. 00:33:13.197 [2024-11-06 09:09:03.061007] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.197 [2024-11-06 09:09:03.061067] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.197 [2024-11-06 09:09:03.061080] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.197 [2024-11-06 09:09:03.061088] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.197 [2024-11-06 09:09:03.061094] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.197 [2024-11-06 09:09:03.061108] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.197 qpair failed and we were unable to recover it. 00:33:13.197 [2024-11-06 09:09:03.071186] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.197 [2024-11-06 09:09:03.071248] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.197 [2024-11-06 09:09:03.071262] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.197 [2024-11-06 09:09:03.071269] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.197 [2024-11-06 09:09:03.071276] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.197 [2024-11-06 09:09:03.071289] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.197 qpair failed and we were unable to recover it. 00:33:13.197 [2024-11-06 09:09:03.081184] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.197 [2024-11-06 09:09:03.081266] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.197 [2024-11-06 09:09:03.081280] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.197 [2024-11-06 09:09:03.081289] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.197 [2024-11-06 09:09:03.081295] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.197 [2024-11-06 09:09:03.081310] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.197 qpair failed and we were unable to recover it. 00:33:13.197 [2024-11-06 09:09:03.091152] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.197 [2024-11-06 09:09:03.091203] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.197 [2024-11-06 09:09:03.091217] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.197 [2024-11-06 09:09:03.091225] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.197 [2024-11-06 09:09:03.091232] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.197 [2024-11-06 09:09:03.091246] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.197 qpair failed and we were unable to recover it. 00:33:13.197 [2024-11-06 09:09:03.101231] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.197 [2024-11-06 09:09:03.101286] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.197 [2024-11-06 09:09:03.101300] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.197 [2024-11-06 09:09:03.101311] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.197 [2024-11-06 09:09:03.101318] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.197 [2024-11-06 09:09:03.101332] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.197 qpair failed and we were unable to recover it. 00:33:13.197 [2024-11-06 09:09:03.111268] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.197 [2024-11-06 09:09:03.111323] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.197 [2024-11-06 09:09:03.111338] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.197 [2024-11-06 09:09:03.111345] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.197 [2024-11-06 09:09:03.111352] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.197 [2024-11-06 09:09:03.111367] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.197 qpair failed and we were unable to recover it. 00:33:13.197 [2024-11-06 09:09:03.121232] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.197 [2024-11-06 09:09:03.121287] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.197 [2024-11-06 09:09:03.121300] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.197 [2024-11-06 09:09:03.121307] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.197 [2024-11-06 09:09:03.121314] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.197 [2024-11-06 09:09:03.121328] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.131264] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.131311] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.131325] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.131332] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.131338] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.131352] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.141341] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.141397] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.141412] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.141419] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.141426] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.141443] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.151400] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.151457] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.151471] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.151479] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.151486] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.151500] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.161383] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.161481] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.161495] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.161502] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.161509] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.161523] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.171372] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.171447] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.171473] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.171482] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.171490] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.171509] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.181442] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.181507] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.181533] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.181542] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.181550] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.181570] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.191473] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.191535] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.191550] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.191559] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.191565] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.191580] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.201340] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.201394] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.201409] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.201417] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.201423] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.201438] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.211495] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.211542] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.211556] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.211563] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.211570] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.211585] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.221561] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.221655] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.221670] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.221677] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.221684] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.221698] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.231587] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.231646] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.231660] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.231672] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.231679] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.231693] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.241572] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.241618] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.241631] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.241639] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.241646] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.241659] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.251587] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.198 [2024-11-06 09:09:03.251643] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.198 [2024-11-06 09:09:03.251656] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.198 [2024-11-06 09:09:03.251664] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.198 [2024-11-06 09:09:03.251670] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.198 [2024-11-06 09:09:03.251684] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.198 qpair failed and we were unable to recover it. 00:33:13.198 [2024-11-06 09:09:03.261649] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.199 [2024-11-06 09:09:03.261716] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.199 [2024-11-06 09:09:03.261729] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.199 [2024-11-06 09:09:03.261737] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.199 [2024-11-06 09:09:03.261743] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.199 [2024-11-06 09:09:03.261763] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.199 qpair failed and we were unable to recover it. 00:33:13.199 [2024-11-06 09:09:03.271698] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.199 [2024-11-06 09:09:03.271762] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.199 [2024-11-06 09:09:03.271776] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.199 [2024-11-06 09:09:03.271784] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.199 [2024-11-06 09:09:03.271790] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.199 [2024-11-06 09:09:03.271808] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.199 qpair failed and we were unable to recover it. 00:33:13.199 [2024-11-06 09:09:03.281685] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.199 [2024-11-06 09:09:03.281729] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.199 [2024-11-06 09:09:03.281743] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.199 [2024-11-06 09:09:03.281754] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.199 [2024-11-06 09:09:03.281761] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.199 [2024-11-06 09:09:03.281775] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.199 qpair failed and we were unable to recover it. 00:33:13.199 [2024-11-06 09:09:03.291710] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.199 [2024-11-06 09:09:03.291762] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.199 [2024-11-06 09:09:03.291776] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.199 [2024-11-06 09:09:03.291784] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.199 [2024-11-06 09:09:03.291790] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.199 [2024-11-06 09:09:03.291804] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.199 qpair failed and we were unable to recover it. 00:33:13.199 [2024-11-06 09:09:03.301778] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.199 [2024-11-06 09:09:03.301839] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.199 [2024-11-06 09:09:03.301853] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.199 [2024-11-06 09:09:03.301860] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.199 [2024-11-06 09:09:03.301866] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.199 [2024-11-06 09:09:03.301880] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.199 qpair failed and we were unable to recover it. 00:33:13.460 [2024-11-06 09:09:03.311728] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.460 [2024-11-06 09:09:03.311824] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.460 [2024-11-06 09:09:03.311839] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.460 [2024-11-06 09:09:03.311846] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.460 [2024-11-06 09:09:03.311853] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.460 [2024-11-06 09:09:03.311867] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.460 qpair failed and we were unable to recover it. 00:33:13.460 [2024-11-06 09:09:03.321787] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.460 [2024-11-06 09:09:03.321837] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.460 [2024-11-06 09:09:03.321852] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.460 [2024-11-06 09:09:03.321859] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.460 [2024-11-06 09:09:03.321866] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.460 [2024-11-06 09:09:03.321881] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.460 qpair failed and we were unable to recover it. 00:33:13.460 [2024-11-06 09:09:03.331859] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.460 [2024-11-06 09:09:03.331930] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.460 [2024-11-06 09:09:03.331944] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.460 [2024-11-06 09:09:03.331951] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.460 [2024-11-06 09:09:03.331958] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.460 [2024-11-06 09:09:03.331973] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.460 qpair failed and we were unable to recover it. 00:33:13.460 [2024-11-06 09:09:03.341813] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.460 [2024-11-06 09:09:03.341884] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.460 [2024-11-06 09:09:03.341899] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.460 [2024-11-06 09:09:03.341906] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.460 [2024-11-06 09:09:03.341912] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.460 [2024-11-06 09:09:03.341928] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.460 qpair failed and we were unable to recover it. 00:33:13.460 [2024-11-06 09:09:03.351922] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.460 [2024-11-06 09:09:03.351973] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.460 [2024-11-06 09:09:03.351987] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.460 [2024-11-06 09:09:03.351995] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.460 [2024-11-06 09:09:03.352001] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.460 [2024-11-06 09:09:03.352015] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.460 qpair failed and we were unable to recover it. 00:33:13.460 [2024-11-06 09:09:03.361896] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.460 [2024-11-06 09:09:03.361946] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.460 [2024-11-06 09:09:03.361960] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.460 [2024-11-06 09:09:03.361971] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.460 [2024-11-06 09:09:03.361978] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.460 [2024-11-06 09:09:03.361992] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.460 qpair failed and we were unable to recover it. 00:33:13.460 [2024-11-06 09:09:03.371923] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.460 [2024-11-06 09:09:03.371974] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.460 [2024-11-06 09:09:03.371987] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.460 [2024-11-06 09:09:03.371994] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.460 [2024-11-06 09:09:03.372001] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.460 [2024-11-06 09:09:03.372015] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.460 qpair failed and we were unable to recover it. 00:33:13.460 [2024-11-06 09:09:03.381994] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.382052] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.382066] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.382074] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.382080] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.382095] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.392043] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.392102] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.392115] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.392123] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.392129] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.392143] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.402018] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.402070] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.402084] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.402092] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.402098] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.402116] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.412076] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.412151] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.412165] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.412173] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.412180] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.412194] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.422112] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.422172] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.422186] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.422193] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.422200] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.422214] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.432140] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.432194] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.432208] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.432215] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.432222] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.432236] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.442090] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.442142] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.442155] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.442163] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.442169] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.442183] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.452014] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.452071] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.452087] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.452095] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.452101] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.452116] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.462216] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.462273] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.462287] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.462294] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.462301] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.462315] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.472127] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.472210] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.472223] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.472231] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.472238] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.472252] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.482213] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.482264] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.482277] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.482285] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.482291] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.482305] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.492260] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.492310] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.492323] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.492337] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.492344] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.492357] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.502331] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.502423] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.502437] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.461 [2024-11-06 09:09:03.502445] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.461 [2024-11-06 09:09:03.502452] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.461 [2024-11-06 09:09:03.502465] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.461 qpair failed and we were unable to recover it. 00:33:13.461 [2024-11-06 09:09:03.512360] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.461 [2024-11-06 09:09:03.512413] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.461 [2024-11-06 09:09:03.512426] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.462 [2024-11-06 09:09:03.512434] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.462 [2024-11-06 09:09:03.512440] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.462 [2024-11-06 09:09:03.512454] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.462 qpair failed and we were unable to recover it. 00:33:13.462 [2024-11-06 09:09:03.522329] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.462 [2024-11-06 09:09:03.522379] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.462 [2024-11-06 09:09:03.522393] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.462 [2024-11-06 09:09:03.522401] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.462 [2024-11-06 09:09:03.522408] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.462 [2024-11-06 09:09:03.522422] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.462 qpair failed and we were unable to recover it. 00:33:13.462 [2024-11-06 09:09:03.532352] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.462 [2024-11-06 09:09:03.532450] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.462 [2024-11-06 09:09:03.532465] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.462 [2024-11-06 09:09:03.532472] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.462 [2024-11-06 09:09:03.532479] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.462 [2024-11-06 09:09:03.532497] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.462 qpair failed and we were unable to recover it. 00:33:13.462 [2024-11-06 09:09:03.542416] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.462 [2024-11-06 09:09:03.542474] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.462 [2024-11-06 09:09:03.542488] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.462 [2024-11-06 09:09:03.542495] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.462 [2024-11-06 09:09:03.542502] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.462 [2024-11-06 09:09:03.542516] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.462 qpair failed and we were unable to recover it. 00:33:13.462 [2024-11-06 09:09:03.552475] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.462 [2024-11-06 09:09:03.552534] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.462 [2024-11-06 09:09:03.552560] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.462 [2024-11-06 09:09:03.552569] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.462 [2024-11-06 09:09:03.552576] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.462 [2024-11-06 09:09:03.552596] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.462 qpair failed and we were unable to recover it. 00:33:13.462 [2024-11-06 09:09:03.562327] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.462 [2024-11-06 09:09:03.562378] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.462 [2024-11-06 09:09:03.562394] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.462 [2024-11-06 09:09:03.562402] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.462 [2024-11-06 09:09:03.562409] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.462 [2024-11-06 09:09:03.562425] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.462 qpair failed and we were unable to recover it. 00:33:13.724 [2024-11-06 09:09:03.572471] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.724 [2024-11-06 09:09:03.572523] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.724 [2024-11-06 09:09:03.572538] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.724 [2024-11-06 09:09:03.572546] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.724 [2024-11-06 09:09:03.572552] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.724 [2024-11-06 09:09:03.572567] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.724 qpair failed and we were unable to recover it. 00:33:13.724 [2024-11-06 09:09:03.582547] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.724 [2024-11-06 09:09:03.582617] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.724 [2024-11-06 09:09:03.582642] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.724 [2024-11-06 09:09:03.582651] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.724 [2024-11-06 09:09:03.582659] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.724 [2024-11-06 09:09:03.582679] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.724 qpair failed and we were unable to recover it. 00:33:13.724 [2024-11-06 09:09:03.592586] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.724 [2024-11-06 09:09:03.592644] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.724 [2024-11-06 09:09:03.592660] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.724 [2024-11-06 09:09:03.592668] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.724 [2024-11-06 09:09:03.592675] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.724 [2024-11-06 09:09:03.592690] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.724 qpair failed and we were unable to recover it. 00:33:13.724 [2024-11-06 09:09:03.602531] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.724 [2024-11-06 09:09:03.602583] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.724 [2024-11-06 09:09:03.602598] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.724 [2024-11-06 09:09:03.602606] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.724 [2024-11-06 09:09:03.602613] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.724 [2024-11-06 09:09:03.602628] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.724 qpair failed and we were unable to recover it. 00:33:13.724 [2024-11-06 09:09:03.612568] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.724 [2024-11-06 09:09:03.612615] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.724 [2024-11-06 09:09:03.612629] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.724 [2024-11-06 09:09:03.612637] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.724 [2024-11-06 09:09:03.612643] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.724 [2024-11-06 09:09:03.612657] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.724 qpair failed and we were unable to recover it. 00:33:13.724 [2024-11-06 09:09:03.622571] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.724 [2024-11-06 09:09:03.622631] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.724 [2024-11-06 09:09:03.622645] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.724 [2024-11-06 09:09:03.622657] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.724 [2024-11-06 09:09:03.622664] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.724 [2024-11-06 09:09:03.622678] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.724 qpair failed and we were unable to recover it. 00:33:13.724 [2024-11-06 09:09:03.632701] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.724 [2024-11-06 09:09:03.632761] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.724 [2024-11-06 09:09:03.632776] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.724 [2024-11-06 09:09:03.632783] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.724 [2024-11-06 09:09:03.632790] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.724 [2024-11-06 09:09:03.632805] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.724 qpair failed and we were unable to recover it. 00:33:13.724 [2024-11-06 09:09:03.642668] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.724 [2024-11-06 09:09:03.642716] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.724 [2024-11-06 09:09:03.642730] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.724 [2024-11-06 09:09:03.642737] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.724 [2024-11-06 09:09:03.642744] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.724 [2024-11-06 09:09:03.642763] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.724 qpair failed and we were unable to recover it. 00:33:13.724 [2024-11-06 09:09:03.652705] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.724 [2024-11-06 09:09:03.652762] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.724 [2024-11-06 09:09:03.652776] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.724 [2024-11-06 09:09:03.652783] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.724 [2024-11-06 09:09:03.652790] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.724 [2024-11-06 09:09:03.652803] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.724 qpair failed and we were unable to recover it. 00:33:13.724 [2024-11-06 09:09:03.662790] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.724 [2024-11-06 09:09:03.662855] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.724 [2024-11-06 09:09:03.662868] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.724 [2024-11-06 09:09:03.662876] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.724 [2024-11-06 09:09:03.662883] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.724 [2024-11-06 09:09:03.662901] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.724 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.672795] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.672849] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.672863] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.672870] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.672877] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.672891] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.682783] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.682831] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.682845] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.682852] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.682859] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.682873] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.692808] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.692861] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.692875] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.692883] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.692889] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.692903] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.702758] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.702813] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.702827] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.702834] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.702841] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.702855] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.712769] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.712832] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.712847] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.712854] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.712860] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.712874] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.722909] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.722959] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.722972] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.722980] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.722986] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.723000] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.732928] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.732982] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.732999] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.733006] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.733013] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.733028] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.742864] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.742922] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.742937] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.742945] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.742951] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.742965] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.752979] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.753028] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.753042] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.753052] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.753059] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.753073] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.762981] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.763033] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.763047] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.763054] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.763061] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.763074] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.773029] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.773081] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.773094] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.773102] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.773108] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.773122] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.783082] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.783139] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.783153] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.783160] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.783167] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.783181] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.725 [2024-11-06 09:09:03.793066] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.725 [2024-11-06 09:09:03.793132] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.725 [2024-11-06 09:09:03.793146] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.725 [2024-11-06 09:09:03.793153] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.725 [2024-11-06 09:09:03.793160] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.725 [2024-11-06 09:09:03.793177] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.725 qpair failed and we were unable to recover it. 00:33:13.726 [2024-11-06 09:09:03.803054] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.726 [2024-11-06 09:09:03.803102] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.726 [2024-11-06 09:09:03.803116] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.726 [2024-11-06 09:09:03.803123] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.726 [2024-11-06 09:09:03.803129] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.726 [2024-11-06 09:09:03.803143] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.726 qpair failed and we were unable to recover it. 00:33:13.726 [2024-11-06 09:09:03.813084] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.726 [2024-11-06 09:09:03.813137] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.726 [2024-11-06 09:09:03.813151] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.726 [2024-11-06 09:09:03.813158] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.726 [2024-11-06 09:09:03.813165] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.726 [2024-11-06 09:09:03.813179] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.726 qpair failed and we were unable to recover it. 00:33:13.726 [2024-11-06 09:09:03.823214] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.726 [2024-11-06 09:09:03.823272] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.726 [2024-11-06 09:09:03.823286] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.726 [2024-11-06 09:09:03.823293] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.726 [2024-11-06 09:09:03.823299] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.726 [2024-11-06 09:09:03.823313] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.726 qpair failed and we were unable to recover it. 00:33:13.726 [2024-11-06 09:09:03.833173] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.726 [2024-11-06 09:09:03.833226] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.726 [2024-11-06 09:09:03.833240] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.726 [2024-11-06 09:09:03.833247] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.726 [2024-11-06 09:09:03.833254] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.726 [2024-11-06 09:09:03.833268] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.726 qpair failed and we were unable to recover it. 00:33:13.988 [2024-11-06 09:09:03.843212] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.988 [2024-11-06 09:09:03.843261] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.988 [2024-11-06 09:09:03.843275] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.988 [2024-11-06 09:09:03.843283] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.988 [2024-11-06 09:09:03.843289] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.988 [2024-11-06 09:09:03.843303] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.988 qpair failed and we were unable to recover it. 00:33:13.988 [2024-11-06 09:09:03.853222] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.988 [2024-11-06 09:09:03.853273] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.988 [2024-11-06 09:09:03.853287] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.988 [2024-11-06 09:09:03.853294] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.988 [2024-11-06 09:09:03.853300] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.988 [2024-11-06 09:09:03.853314] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.988 qpair failed and we were unable to recover it. 00:33:13.988 [2024-11-06 09:09:03.863214] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.988 [2024-11-06 09:09:03.863270] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.988 [2024-11-06 09:09:03.863283] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.988 [2024-11-06 09:09:03.863291] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.988 [2024-11-06 09:09:03.863297] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.988 [2024-11-06 09:09:03.863311] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.988 qpair failed and we were unable to recover it. 00:33:13.988 [2024-11-06 09:09:03.873300] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.988 [2024-11-06 09:09:03.873352] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.988 [2024-11-06 09:09:03.873365] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.988 [2024-11-06 09:09:03.873373] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.988 [2024-11-06 09:09:03.873379] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.988 [2024-11-06 09:09:03.873394] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.988 qpair failed and we were unable to recover it. 00:33:13.988 [2024-11-06 09:09:03.883314] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.988 [2024-11-06 09:09:03.883363] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.988 [2024-11-06 09:09:03.883376] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.988 [2024-11-06 09:09:03.883387] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.988 [2024-11-06 09:09:03.883394] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.988 [2024-11-06 09:09:03.883408] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.988 qpair failed and we were unable to recover it. 00:33:13.988 [2024-11-06 09:09:03.893348] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.988 [2024-11-06 09:09:03.893404] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.988 [2024-11-06 09:09:03.893418] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.988 [2024-11-06 09:09:03.893426] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.988 [2024-11-06 09:09:03.893432] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.988 [2024-11-06 09:09:03.893446] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.988 qpair failed and we were unable to recover it. 00:33:13.988 [2024-11-06 09:09:03.903422] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.988 [2024-11-06 09:09:03.903478] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.988 [2024-11-06 09:09:03.903492] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.988 [2024-11-06 09:09:03.903500] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.988 [2024-11-06 09:09:03.903506] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.988 [2024-11-06 09:09:03.903520] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.988 qpair failed and we were unable to recover it. 00:33:13.988 [2024-11-06 09:09:03.913365] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.988 [2024-11-06 09:09:03.913416] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.988 [2024-11-06 09:09:03.913431] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.988 [2024-11-06 09:09:03.913439] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.988 [2024-11-06 09:09:03.913445] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.988 [2024-11-06 09:09:03.913459] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.988 qpair failed and we were unable to recover it. 00:33:13.988 [2024-11-06 09:09:03.923421] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.988 [2024-11-06 09:09:03.923471] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.988 [2024-11-06 09:09:03.923485] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:03.923492] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:03.923499] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:03.923516] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:03.933463] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:03.933514] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:03.933528] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:03.933535] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:03.933542] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:03.933556] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:03.943534] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:03.943593] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:03.943611] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:03.943619] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:03.943625] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:03.943640] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:03.953518] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:03.953576] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:03.953602] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:03.953610] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:03.953618] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:03.953638] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:03.963545] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:03.963593] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:03.963609] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:03.963616] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:03.963623] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:03.963638] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:03.973576] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:03.973630] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:03.973644] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:03.973652] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:03.973658] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:03.973673] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:03.983613] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:03.983675] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:03.983689] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:03.983696] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:03.983703] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:03.983717] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:03.993606] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:03.993661] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:03.993674] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:03.993682] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:03.993688] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:03.993702] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:04.003637] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:04.003691] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:04.003705] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:04.003713] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:04.003720] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:04.003734] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:04.013681] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:04.013730] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:04.013744] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:04.013760] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:04.013767] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:04.013781] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:04.023752] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:04.023811] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:04.023825] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:04.023833] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:04.023840] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:04.023854] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:04.033740] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:04.033795] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:04.033808] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:04.033816] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:04.033822] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:04.033836] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:04.043753] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:04.043808] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:04.043821] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:04.043829] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:04.043835] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.989 [2024-11-06 09:09:04.043849] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.989 qpair failed and we were unable to recover it. 00:33:13.989 [2024-11-06 09:09:04.053761] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.989 [2024-11-06 09:09:04.053813] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.989 [2024-11-06 09:09:04.053826] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.989 [2024-11-06 09:09:04.053834] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.989 [2024-11-06 09:09:04.053840] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.990 [2024-11-06 09:09:04.053857] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.990 qpair failed and we were unable to recover it. 00:33:13.990 [2024-11-06 09:09:04.063848] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.990 [2024-11-06 09:09:04.063905] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.990 [2024-11-06 09:09:04.063919] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.990 [2024-11-06 09:09:04.063926] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.990 [2024-11-06 09:09:04.063933] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.990 [2024-11-06 09:09:04.063947] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.990 qpair failed and we were unable to recover it. 00:33:13.990 [2024-11-06 09:09:04.073879] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.990 [2024-11-06 09:09:04.073937] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.990 [2024-11-06 09:09:04.073950] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.990 [2024-11-06 09:09:04.073957] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.990 [2024-11-06 09:09:04.073963] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.990 [2024-11-06 09:09:04.073977] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.990 qpair failed and we were unable to recover it. 00:33:13.990 [2024-11-06 09:09:04.083836] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.990 [2024-11-06 09:09:04.083891] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.990 [2024-11-06 09:09:04.083904] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.990 [2024-11-06 09:09:04.083912] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.990 [2024-11-06 09:09:04.083918] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.990 [2024-11-06 09:09:04.083932] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.990 qpair failed and we were unable to recover it. 00:33:13.990 [2024-11-06 09:09:04.093887] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:13.990 [2024-11-06 09:09:04.093942] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:13.990 [2024-11-06 09:09:04.093956] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:13.990 [2024-11-06 09:09:04.093964] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:13.990 [2024-11-06 09:09:04.093970] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:13.990 [2024-11-06 09:09:04.093984] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:13.990 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.103861] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.103922] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.103936] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.103943] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.103950] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.103964] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.113953] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.114002] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.114015] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.114023] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.114029] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.114043] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.123969] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.124023] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.124037] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.124045] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.124051] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.124066] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.133992] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.134058] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.134072] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.134079] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.134085] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.134100] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.144080] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.144139] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.144153] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.144164] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.144170] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.144184] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.154043] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.154092] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.154107] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.154114] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.154121] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.154135] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.164083] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.164130] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.164145] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.164152] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.164159] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.164173] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.174065] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.174110] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.174124] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.174131] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.174138] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.174152] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.184108] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.184171] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.184185] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.184192] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.184199] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.184212] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.194160] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.194214] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.194228] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.194235] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.194241] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.194255] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.204174] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.204224] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.204237] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.204245] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.204251] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.204266] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.214195] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.214248] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.214262] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.253 [2024-11-06 09:09:04.214269] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.253 [2024-11-06 09:09:04.214276] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.253 [2024-11-06 09:09:04.214289] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.253 qpair failed and we were unable to recover it. 00:33:14.253 [2024-11-06 09:09:04.224241] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.253 [2024-11-06 09:09:04.224299] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.253 [2024-11-06 09:09:04.224312] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.224320] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.224326] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.224340] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.234267] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.234317] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.234331] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.234338] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.234345] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.234358] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.244292] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.244337] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.244351] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.244359] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.244365] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.244379] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.254170] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.254246] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.254260] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.254267] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.254274] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.254288] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.264328] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.264373] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.264386] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.264394] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.264401] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.264415] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.274375] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.274465] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.274479] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.274491] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.274497] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.274511] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.284392] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.284444] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.284458] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.284466] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.284472] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.284486] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.294407] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.294449] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.294462] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.294470] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.294476] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.294490] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.304421] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.304469] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.304482] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.304490] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.304496] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.304510] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.314345] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.314392] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.314406] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.314414] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.314420] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.314434] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.324485] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.324535] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.324549] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.324556] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.324563] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.324577] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.334522] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.334569] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.334582] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.334589] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.334596] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.334610] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.344549] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.344607] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.344634] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.344643] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.254 [2024-11-06 09:09:04.344651] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.254 [2024-11-06 09:09:04.344671] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.254 qpair failed and we were unable to recover it. 00:33:14.254 [2024-11-06 09:09:04.354456] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.254 [2024-11-06 09:09:04.354505] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.254 [2024-11-06 09:09:04.354521] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.254 [2024-11-06 09:09:04.354528] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.255 [2024-11-06 09:09:04.354535] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.255 [2024-11-06 09:09:04.354550] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.255 qpair failed and we were unable to recover it. 00:33:14.517 [2024-11-06 09:09:04.364604] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.517 [2024-11-06 09:09:04.364653] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.517 [2024-11-06 09:09:04.364668] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.517 [2024-11-06 09:09:04.364676] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.517 [2024-11-06 09:09:04.364682] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.517 [2024-11-06 09:09:04.364697] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.517 qpair failed and we were unable to recover it. 00:33:14.517 [2024-11-06 09:09:04.374619] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.517 [2024-11-06 09:09:04.374701] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.517 [2024-11-06 09:09:04.374715] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.517 [2024-11-06 09:09:04.374722] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.517 [2024-11-06 09:09:04.374729] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.517 [2024-11-06 09:09:04.374743] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.517 qpair failed and we were unable to recover it. 00:33:14.517 [2024-11-06 09:09:04.384666] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.517 [2024-11-06 09:09:04.384715] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.517 [2024-11-06 09:09:04.384729] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.517 [2024-11-06 09:09:04.384737] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.517 [2024-11-06 09:09:04.384744] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.517 [2024-11-06 09:09:04.384763] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.517 qpair failed and we were unable to recover it. 00:33:14.517 [2024-11-06 09:09:04.394711] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.517 [2024-11-06 09:09:04.394765] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.517 [2024-11-06 09:09:04.394779] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.517 [2024-11-06 09:09:04.394787] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.394794] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.394808] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.404675] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.404720] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.404735] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.404750] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.404757] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.404771] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.414723] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.414769] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.414783] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.414790] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.414796] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.414811] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.424736] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.424789] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.424803] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.424810] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.424817] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.424831] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.434778] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.434828] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.434842] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.434849] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.434856] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.434870] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.444810] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.444857] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.444870] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.444878] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.444884] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.444900] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.454848] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.454897] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.454910] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.454917] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.454924] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.454938] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.464765] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.464815] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.464828] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.464835] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.464842] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.464856] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.474930] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.474983] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.474997] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.475004] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.475010] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.475024] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.484927] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.485002] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.485017] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.485024] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.485030] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.485044] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.494967] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.495029] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.495043] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.495050] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.495057] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.495070] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.504987] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.505033] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.505047] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.505054] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.505061] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.505075] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.514894] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.514947] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.514960] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.514968] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.514974] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.518 [2024-11-06 09:09:04.514988] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.518 qpair failed and we were unable to recover it. 00:33:14.518 [2024-11-06 09:09:04.525034] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.518 [2024-11-06 09:09:04.525082] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.518 [2024-11-06 09:09:04.525095] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.518 [2024-11-06 09:09:04.525103] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.518 [2024-11-06 09:09:04.525110] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.525123] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.519 [2024-11-06 09:09:04.535068] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.519 [2024-11-06 09:09:04.535144] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.519 [2024-11-06 09:09:04.535157] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.519 [2024-11-06 09:09:04.535169] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.519 [2024-11-06 09:09:04.535177] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.535191] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.519 [2024-11-06 09:09:04.545079] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.519 [2024-11-06 09:09:04.545137] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.519 [2024-11-06 09:09:04.545150] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.519 [2024-11-06 09:09:04.545158] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.519 [2024-11-06 09:09:04.545164] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.545178] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.519 [2024-11-06 09:09:04.555132] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.519 [2024-11-06 09:09:04.555180] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.519 [2024-11-06 09:09:04.555194] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.519 [2024-11-06 09:09:04.555201] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.519 [2024-11-06 09:09:04.555207] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.555221] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.519 [2024-11-06 09:09:04.565151] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.519 [2024-11-06 09:09:04.565198] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.519 [2024-11-06 09:09:04.565212] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.519 [2024-11-06 09:09:04.565220] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.519 [2024-11-06 09:09:04.565226] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.565241] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.519 [2024-11-06 09:09:04.575052] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.519 [2024-11-06 09:09:04.575100] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.519 [2024-11-06 09:09:04.575114] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.519 [2024-11-06 09:09:04.575121] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.519 [2024-11-06 09:09:04.575127] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.575142] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.519 [2024-11-06 09:09:04.585192] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.519 [2024-11-06 09:09:04.585271] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.519 [2024-11-06 09:09:04.585285] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.519 [2024-11-06 09:09:04.585293] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.519 [2024-11-06 09:09:04.585300] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.585314] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.519 [2024-11-06 09:09:04.595233] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.519 [2024-11-06 09:09:04.595283] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.519 [2024-11-06 09:09:04.595297] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.519 [2024-11-06 09:09:04.595304] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.519 [2024-11-06 09:09:04.595311] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.595325] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.519 [2024-11-06 09:09:04.605249] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.519 [2024-11-06 09:09:04.605297] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.519 [2024-11-06 09:09:04.605311] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.519 [2024-11-06 09:09:04.605318] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.519 [2024-11-06 09:09:04.605324] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.605338] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.519 [2024-11-06 09:09:04.615336] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.519 [2024-11-06 09:09:04.615392] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.519 [2024-11-06 09:09:04.615405] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.519 [2024-11-06 09:09:04.615412] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.519 [2024-11-06 09:09:04.615419] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.615433] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.519 [2024-11-06 09:09:04.625393] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.519 [2024-11-06 09:09:04.625466] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.519 [2024-11-06 09:09:04.625480] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.519 [2024-11-06 09:09:04.625487] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.519 [2024-11-06 09:09:04.625494] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.519 [2024-11-06 09:09:04.625508] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.519 qpair failed and we were unable to recover it. 00:33:14.782 [2024-11-06 09:09:04.635342] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.782 [2024-11-06 09:09:04.635392] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.782 [2024-11-06 09:09:04.635406] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.782 [2024-11-06 09:09:04.635413] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.782 [2024-11-06 09:09:04.635420] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.782 [2024-11-06 09:09:04.635434] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.782 qpair failed and we were unable to recover it. 00:33:14.782 [2024-11-06 09:09:04.645402] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.782 [2024-11-06 09:09:04.645450] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.782 [2024-11-06 09:09:04.645464] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.782 [2024-11-06 09:09:04.645472] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.782 [2024-11-06 09:09:04.645478] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.782 [2024-11-06 09:09:04.645492] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.782 qpair failed and we were unable to recover it. 00:33:14.782 [2024-11-06 09:09:04.655367] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.782 [2024-11-06 09:09:04.655414] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.782 [2024-11-06 09:09:04.655428] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.782 [2024-11-06 09:09:04.655435] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.782 [2024-11-06 09:09:04.655442] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.782 [2024-11-06 09:09:04.655456] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.782 qpair failed and we were unable to recover it. 00:33:14.782 [2024-11-06 09:09:04.665409] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.782 [2024-11-06 09:09:04.665461] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.782 [2024-11-06 09:09:04.665475] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.782 [2024-11-06 09:09:04.665485] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.782 [2024-11-06 09:09:04.665492] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.782 [2024-11-06 09:09:04.665506] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.782 qpair failed and we were unable to recover it. 00:33:14.782 [2024-11-06 09:09:04.675451] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.782 [2024-11-06 09:09:04.675522] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.782 [2024-11-06 09:09:04.675536] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.782 [2024-11-06 09:09:04.675543] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.675550] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.675564] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.685457] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.685507] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.685521] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.685528] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.685534] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.685548] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.695483] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.695551] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.695577] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.695586] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.695594] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.695613] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.705515] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.705569] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.705586] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.705593] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.705600] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.705615] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.715565] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.715612] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.715626] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.715634] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.715640] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.715655] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.725549] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.725595] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.725609] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.725616] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.725623] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.725637] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.735613] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.735661] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.735677] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.735686] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.735693] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.735707] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.745626] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.745680] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.745693] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.745701] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.745708] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.745722] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.755663] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.755716] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.755729] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.755737] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.755743] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.755763] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.765608] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.765657] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.765671] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.765679] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.765685] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.765699] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.775708] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.775766] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.775783] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.775791] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.775798] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.783 [2024-11-06 09:09:04.775814] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.783 qpair failed and we were unable to recover it. 00:33:14.783 [2024-11-06 09:09:04.785752] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.783 [2024-11-06 09:09:04.785800] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.783 [2024-11-06 09:09:04.785814] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.783 [2024-11-06 09:09:04.785821] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.783 [2024-11-06 09:09:04.785828] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.785842] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:14.784 [2024-11-06 09:09:04.795752] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.784 [2024-11-06 09:09:04.795803] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.784 [2024-11-06 09:09:04.795818] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.784 [2024-11-06 09:09:04.795830] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.784 [2024-11-06 09:09:04.795837] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.795851] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:14.784 [2024-11-06 09:09:04.805797] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.784 [2024-11-06 09:09:04.805843] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.784 [2024-11-06 09:09:04.805857] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.784 [2024-11-06 09:09:04.805864] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.784 [2024-11-06 09:09:04.805871] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.805885] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:14.784 [2024-11-06 09:09:04.815824] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.784 [2024-11-06 09:09:04.815869] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.784 [2024-11-06 09:09:04.815883] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.784 [2024-11-06 09:09:04.815890] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.784 [2024-11-06 09:09:04.815896] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.815911] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:14.784 [2024-11-06 09:09:04.825828] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.784 [2024-11-06 09:09:04.825881] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.784 [2024-11-06 09:09:04.825895] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.784 [2024-11-06 09:09:04.825902] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.784 [2024-11-06 09:09:04.825909] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.825923] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:14.784 [2024-11-06 09:09:04.835752] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.784 [2024-11-06 09:09:04.835803] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.784 [2024-11-06 09:09:04.835816] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.784 [2024-11-06 09:09:04.835824] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.784 [2024-11-06 09:09:04.835830] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.835844] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:14.784 [2024-11-06 09:09:04.845895] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.784 [2024-11-06 09:09:04.845939] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.784 [2024-11-06 09:09:04.845953] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.784 [2024-11-06 09:09:04.845960] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.784 [2024-11-06 09:09:04.845967] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.845981] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:14.784 [2024-11-06 09:09:04.855886] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.784 [2024-11-06 09:09:04.855932] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.784 [2024-11-06 09:09:04.855946] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.784 [2024-11-06 09:09:04.855954] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.784 [2024-11-06 09:09:04.855960] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.855974] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:14.784 [2024-11-06 09:09:04.865938] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.784 [2024-11-06 09:09:04.865991] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.784 [2024-11-06 09:09:04.866005] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.784 [2024-11-06 09:09:04.866012] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.784 [2024-11-06 09:09:04.866018] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.866032] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:14.784 [2024-11-06 09:09:04.875979] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.784 [2024-11-06 09:09:04.876028] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.784 [2024-11-06 09:09:04.876042] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.784 [2024-11-06 09:09:04.876050] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.784 [2024-11-06 09:09:04.876057] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.876071] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:14.784 [2024-11-06 09:09:04.885975] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:14.784 [2024-11-06 09:09:04.886032] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:14.784 [2024-11-06 09:09:04.886046] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:14.784 [2024-11-06 09:09:04.886054] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:14.784 [2024-11-06 09:09:04.886060] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:14.784 [2024-11-06 09:09:04.886074] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:14.784 qpair failed and we were unable to recover it. 00:33:15.047 [2024-11-06 09:09:04.896039] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.047 [2024-11-06 09:09:04.896086] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.047 [2024-11-06 09:09:04.896099] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.047 [2024-11-06 09:09:04.896106] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.047 [2024-11-06 09:09:04.896113] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.047 [2024-11-06 09:09:04.896127] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.047 qpair failed and we were unable to recover it. 00:33:15.047 [2024-11-06 09:09:04.906066] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.047 [2024-11-06 09:09:04.906116] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.047 [2024-11-06 09:09:04.906131] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.047 [2024-11-06 09:09:04.906139] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.047 [2024-11-06 09:09:04.906145] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.047 [2024-11-06 09:09:04.906160] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.047 qpair failed and we were unable to recover it. 00:33:15.047 [2024-11-06 09:09:04.915963] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.047 [2024-11-06 09:09:04.916011] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.047 [2024-11-06 09:09:04.916026] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.047 [2024-11-06 09:09:04.916034] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.047 [2024-11-06 09:09:04.916040] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.047 [2024-11-06 09:09:04.916055] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.047 qpair failed and we were unable to recover it. 00:33:15.047 [2024-11-06 09:09:04.926053] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.047 [2024-11-06 09:09:04.926149] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:04.926164] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:04.926176] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:04.926182] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:04.926197] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:04.936114] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:04.936159] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:04.936173] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:04.936180] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:04.936187] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:04.936201] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:04.946168] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:04.946262] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:04.946277] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:04.946284] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:04.946291] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:04.946305] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:04.956200] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:04.956245] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:04.956259] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:04.956266] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:04.956273] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:04.956287] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:04.966219] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:04.966274] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:04.966287] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:04.966294] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:04.966301] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:04.966314] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:04.976235] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:04.976279] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:04.976294] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:04.976301] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:04.976307] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:04.976321] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:04.986258] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:04.986320] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:04.986335] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:04.986342] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:04.986349] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:04.986363] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:04.996291] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:04.996341] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:04.996355] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:04.996362] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:04.996369] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:04.996382] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:05.006208] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:05.006275] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:05.006289] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:05.006297] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:05.006303] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:05.006318] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:05.016345] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:05.016395] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:05.016409] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:05.016416] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:05.016423] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:05.016436] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:05.026369] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:05.026418] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:05.026432] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:05.026439] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:05.026445] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:05.026459] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:05.036384] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:05.036430] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:05.036443] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:05.036451] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:05.036457] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:05.036471] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:05.046429] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:05.046472] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:05.046486] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:05.046493] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.048 [2024-11-06 09:09:05.046500] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.048 [2024-11-06 09:09:05.046514] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.048 qpair failed and we were unable to recover it. 00:33:15.048 [2024-11-06 09:09:05.056431] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.048 [2024-11-06 09:09:05.056483] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.048 [2024-11-06 09:09:05.056509] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.048 [2024-11-06 09:09:05.056522] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.056530] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.049 [2024-11-06 09:09:05.056550] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 [2024-11-06 09:09:05.066460] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.049 [2024-11-06 09:09:05.066517] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.049 [2024-11-06 09:09:05.066542] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.049 [2024-11-06 09:09:05.066551] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.066559] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.049 [2024-11-06 09:09:05.066578] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 [2024-11-06 09:09:05.076508] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.049 [2024-11-06 09:09:05.076561] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.049 [2024-11-06 09:09:05.076587] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.049 [2024-11-06 09:09:05.076596] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.076603] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.049 [2024-11-06 09:09:05.076622] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 [2024-11-06 09:09:05.086535] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.049 [2024-11-06 09:09:05.086578] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.049 [2024-11-06 09:09:05.086594] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.049 [2024-11-06 09:09:05.086602] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.086608] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.049 [2024-11-06 09:09:05.086623] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 [2024-11-06 09:09:05.096551] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.049 [2024-11-06 09:09:05.096595] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.049 [2024-11-06 09:09:05.096609] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.049 [2024-11-06 09:09:05.096616] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.096623] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.049 [2024-11-06 09:09:05.096637] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 [2024-11-06 09:09:05.106583] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.049 [2024-11-06 09:09:05.106628] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.049 [2024-11-06 09:09:05.106643] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.049 [2024-11-06 09:09:05.106650] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.106656] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.049 [2024-11-06 09:09:05.106671] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 [2024-11-06 09:09:05.116613] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.049 [2024-11-06 09:09:05.116660] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.049 [2024-11-06 09:09:05.116674] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.049 [2024-11-06 09:09:05.116681] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.116687] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.049 [2024-11-06 09:09:05.116701] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 [2024-11-06 09:09:05.126593] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.049 [2024-11-06 09:09:05.126640] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.049 [2024-11-06 09:09:05.126653] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.049 [2024-11-06 09:09:05.126660] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.126667] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.049 [2024-11-06 09:09:05.126681] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 [2024-11-06 09:09:05.136652] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.049 [2024-11-06 09:09:05.136708] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.049 [2024-11-06 09:09:05.136722] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.049 [2024-11-06 09:09:05.136729] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.136735] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x8500c0 00:33:15.049 [2024-11-06 09:09:05.136752] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 3 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 [2024-11-06 09:09:05.146693] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.049 [2024-11-06 09:09:05.146804] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.049 [2024-11-06 09:09:05.146898] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.049 [2024-11-06 09:09:05.146924] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.146944] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f6088000b90 00:33:15.049 [2024-11-06 09:09:05.147002] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 [2024-11-06 09:09:05.156689] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.049 [2024-11-06 09:09:05.156782] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.049 [2024-11-06 09:09:05.156813] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.049 [2024-11-06 09:09:05.156829] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.049 [2024-11-06 09:09:05.156843] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f6088000b90 00:33:15.049 [2024-11-06 09:09:05.156875] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 4 00:33:15.049 qpair failed and we were unable to recover it. 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Write completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Write completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Write completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Write completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Write completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.049 Read completed with error (sct=0, sc=8) 00:33:15.049 starting I/O failed 00:33:15.050 Read completed with error (sct=0, sc=8) 00:33:15.050 starting I/O failed 00:33:15.050 Write completed with error (sct=0, sc=8) 00:33:15.050 starting I/O failed 00:33:15.050 Read completed with error (sct=0, sc=8) 00:33:15.050 starting I/O failed 00:33:15.050 Write completed with error (sct=0, sc=8) 00:33:15.050 starting I/O failed 00:33:15.050 Write completed with error (sct=0, sc=8) 00:33:15.050 starting I/O failed 00:33:15.050 Write completed with error (sct=0, sc=8) 00:33:15.050 starting I/O failed 00:33:15.050 [2024-11-06 09:09:05.157206] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:33:15.310 [2024-11-06 09:09:05.166741] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.310 [2024-11-06 09:09:05.166788] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.310 [2024-11-06 09:09:05.166806] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.310 [2024-11-06 09:09:05.166812] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.310 [2024-11-06 09:09:05.166817] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f608c000b90 00:33:15.310 [2024-11-06 09:09:05.166829] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:33:15.310 qpair failed and we were unable to recover it. 00:33:15.310 [2024-11-06 09:09:05.176777] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.310 [2024-11-06 09:09:05.176825] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.310 [2024-11-06 09:09:05.176844] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.310 [2024-11-06 09:09:05.176851] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.310 [2024-11-06 09:09:05.176856] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f608c000b90 00:33:15.310 [2024-11-06 09:09:05.176869] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 2 00:33:15.310 qpair failed and we were unable to recover it. 00:33:15.310 [2024-11-06 09:09:05.186794] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.310 [2024-11-06 09:09:05.186911] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.310 [2024-11-06 09:09:05.186969] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.310 [2024-11-06 09:09:05.186993] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.310 [2024-11-06 09:09:05.187013] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f6094000b90 00:33:15.310 [2024-11-06 09:09:05.187062] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:33:15.310 qpair failed and we were unable to recover it. 00:33:15.310 [2024-11-06 09:09:05.196774] ctrlr.c: 762:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:33:15.310 [2024-11-06 09:09:05.196856] nvme_fabric.c: 599:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:33:15.310 [2024-11-06 09:09:05.196885] nvme_fabric.c: 610:nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:33:15.310 [2024-11-06 09:09:05.196899] nvme_tcp.c:2348:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:33:15.310 [2024-11-06 09:09:05.196913] nvme_tcp.c:2125:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7f6094000b90 00:33:15.310 [2024-11-06 09:09:05.196946] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] CQ transport error -6 (No such device or address) on qpair id 1 00:33:15.310 qpair failed and we were unable to recover it. 00:33:15.311 [2024-11-06 09:09:05.197097] nvme_ctrlr.c:4482:nvme_ctrlr_keep_alive: *ERROR*: [nqn.2016-06.io.spdk:cnode1, 1] Submitting Keep Alive failed 00:33:15.311 A controller has encountered a failure and is being reset. 00:33:15.311 [2024-11-06 09:09:05.197216] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x845e00 (9): Bad file descriptor 00:33:15.311 Controller properly reset. 00:33:15.311 Initializing NVMe Controllers 00:33:15.311 Attaching to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:33:15.311 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:33:15.311 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 0 00:33:15.311 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 1 00:33:15.311 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 2 00:33:15.311 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 3 00:33:15.311 Initialization complete. Launching workers. 00:33:15.311 Starting thread on core 1 00:33:15.311 Starting thread on core 2 00:33:15.311 Starting thread on core 3 00:33:15.311 Starting thread on core 0 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@51 -- # sync 00:33:15.311 00:33:15.311 real 0m11.530s 00:33:15.311 user 0m21.749s 00:33:15.311 sys 0m3.708s 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:33:15.311 ************************************ 00:33:15.311 END TEST nvmf_target_disconnect_tc2 00:33:15.311 ************************************ 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@72 -- # '[' -n '' ']' 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@76 -- # trap - SIGINT SIGTERM EXIT 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- host/target_disconnect.sh@77 -- # nvmftestfini 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@514 -- # nvmfcleanup 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@121 -- # sync 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@124 -- # set +e 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@125 -- # for i in {1..20} 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:33:15.311 rmmod nvme_tcp 00:33:15.311 rmmod nvme_fabrics 00:33:15.311 rmmod nvme_keyring 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@128 -- # set -e 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@129 -- # return 0 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@515 -- # '[' -n 2081059 ']' 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@516 -- # killprocess 2081059 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@950 -- # '[' -z 2081059 ']' 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@954 -- # kill -0 2081059 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@955 -- # uname 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2081059 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@956 -- # process_name=reactor_4 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@960 -- # '[' reactor_4 = sudo ']' 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2081059' 00:33:15.311 killing process with pid 2081059 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@969 -- # kill 2081059 00:33:15.311 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@974 -- # wait 2081059 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@297 -- # iptr 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@789 -- # iptables-save 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@789 -- # iptables-restore 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@302 -- # remove_spdk_ns 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:15.572 09:09:05 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:17.487 09:09:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:33:17.487 00:33:17.487 real 0m21.831s 00:33:17.487 user 0m49.852s 00:33:17.487 sys 0m9.802s 00:33:17.487 09:09:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:17.487 09:09:07 nvmf_tcp.nvmf_host.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:33:17.487 ************************************ 00:33:17.487 END TEST nvmf_target_disconnect 00:33:17.487 ************************************ 00:33:17.748 09:09:07 nvmf_tcp.nvmf_host -- nvmf/nvmf_host.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:33:17.748 00:33:17.748 real 6m28.730s 00:33:17.748 user 11m26.752s 00:33:17.748 sys 2m10.629s 00:33:17.748 09:09:07 nvmf_tcp.nvmf_host -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:17.748 09:09:07 nvmf_tcp.nvmf_host -- common/autotest_common.sh@10 -- # set +x 00:33:17.748 ************************************ 00:33:17.748 END TEST nvmf_host 00:33:17.748 ************************************ 00:33:17.748 09:09:07 nvmf_tcp -- nvmf/nvmf.sh@19 -- # [[ tcp = \t\c\p ]] 00:33:17.748 09:09:07 nvmf_tcp -- nvmf/nvmf.sh@19 -- # [[ 0 -eq 0 ]] 00:33:17.748 09:09:07 nvmf_tcp -- nvmf/nvmf.sh@20 -- # run_test nvmf_target_core_interrupt_mode /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp --interrupt-mode 00:33:17.748 09:09:07 nvmf_tcp -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:33:17.748 09:09:07 nvmf_tcp -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:17.748 09:09:07 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:33:17.748 ************************************ 00:33:17.748 START TEST nvmf_target_core_interrupt_mode 00:33:17.748 ************************************ 00:33:17.748 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf_target_core.sh --transport=tcp --interrupt-mode 00:33:17.748 * Looking for test storage... 00:33:17.748 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:33:17.748 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:33:17.748 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1689 -- # lcov --version 00:33:17.748 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@336 -- # IFS=.-: 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@336 -- # read -ra ver1 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@337 -- # IFS=.-: 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@337 -- # read -ra ver2 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@338 -- # local 'op=<' 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@340 -- # ver1_l=2 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@341 -- # ver2_l=1 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@344 -- # case "$op" in 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@345 -- # : 1 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@365 -- # decimal 1 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@353 -- # local d=1 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@355 -- # echo 1 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@365 -- # ver1[v]=1 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@366 -- # decimal 2 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@353 -- # local d=2 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@355 -- # echo 2 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@366 -- # ver2[v]=2 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@368 -- # return 0 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:33:18.017 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:18.017 --rc genhtml_branch_coverage=1 00:33:18.017 --rc genhtml_function_coverage=1 00:33:18.017 --rc genhtml_legend=1 00:33:18.017 --rc geninfo_all_blocks=1 00:33:18.017 --rc geninfo_unexecuted_blocks=1 00:33:18.017 00:33:18.017 ' 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:33:18.017 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:18.017 --rc genhtml_branch_coverage=1 00:33:18.017 --rc genhtml_function_coverage=1 00:33:18.017 --rc genhtml_legend=1 00:33:18.017 --rc geninfo_all_blocks=1 00:33:18.017 --rc geninfo_unexecuted_blocks=1 00:33:18.017 00:33:18.017 ' 00:33:18.017 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:33:18.017 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:18.017 --rc genhtml_branch_coverage=1 00:33:18.018 --rc genhtml_function_coverage=1 00:33:18.018 --rc genhtml_legend=1 00:33:18.018 --rc geninfo_all_blocks=1 00:33:18.018 --rc geninfo_unexecuted_blocks=1 00:33:18.018 00:33:18.018 ' 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:33:18.018 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:18.018 --rc genhtml_branch_coverage=1 00:33:18.018 --rc genhtml_function_coverage=1 00:33:18.018 --rc genhtml_legend=1 00:33:18.018 --rc geninfo_all_blocks=1 00:33:18.018 --rc geninfo_unexecuted_blocks=1 00:33:18.018 00:33:18.018 ' 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@10 -- # uname -s 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@10 -- # '[' '!' Linux = Linux ']' 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@7 -- # uname -s 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@15 -- # shopt -s extglob 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@5 -- # export PATH 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@51 -- # : 0 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/common.sh@55 -- # have_pci_nics=0 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@16 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@18 -- # TEST_ARGS=("$@") 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@20 -- # [[ 0 -eq 0 ]] 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@21 -- # run_test nvmf_abort /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp --interrupt-mode 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:33:18.018 ************************************ 00:33:18.018 START TEST nvmf_abort 00:33:18.018 ************************************ 00:33:18.018 09:09:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp --interrupt-mode 00:33:18.018 * Looking for test storage... 00:33:18.018 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:33:18.018 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:33:18.018 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1689 -- # lcov --version 00:33:18.018 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@336 -- # IFS=.-: 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@336 -- # read -ra ver1 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@337 -- # IFS=.-: 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@337 -- # read -ra ver2 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@338 -- # local 'op=<' 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@340 -- # ver1_l=2 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@341 -- # ver2_l=1 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@344 -- # case "$op" in 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@345 -- # : 1 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@365 -- # decimal 1 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@353 -- # local d=1 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@355 -- # echo 1 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@365 -- # ver1[v]=1 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@366 -- # decimal 2 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@353 -- # local d=2 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@355 -- # echo 2 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@366 -- # ver2[v]=2 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@368 -- # return 0 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:33:18.338 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:18.338 --rc genhtml_branch_coverage=1 00:33:18.338 --rc genhtml_function_coverage=1 00:33:18.338 --rc genhtml_legend=1 00:33:18.338 --rc geninfo_all_blocks=1 00:33:18.338 --rc geninfo_unexecuted_blocks=1 00:33:18.338 00:33:18.338 ' 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:33:18.338 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:18.338 --rc genhtml_branch_coverage=1 00:33:18.338 --rc genhtml_function_coverage=1 00:33:18.338 --rc genhtml_legend=1 00:33:18.338 --rc geninfo_all_blocks=1 00:33:18.338 --rc geninfo_unexecuted_blocks=1 00:33:18.338 00:33:18.338 ' 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:33:18.338 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:18.338 --rc genhtml_branch_coverage=1 00:33:18.338 --rc genhtml_function_coverage=1 00:33:18.338 --rc genhtml_legend=1 00:33:18.338 --rc geninfo_all_blocks=1 00:33:18.338 --rc geninfo_unexecuted_blocks=1 00:33:18.338 00:33:18.338 ' 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:33:18.338 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:18.338 --rc genhtml_branch_coverage=1 00:33:18.338 --rc genhtml_function_coverage=1 00:33:18.338 --rc genhtml_legend=1 00:33:18.338 --rc geninfo_all_blocks=1 00:33:18.338 --rc geninfo_unexecuted_blocks=1 00:33:18.338 00:33:18.338 ' 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@7 -- # uname -s 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:33:18.338 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@15 -- # shopt -s extglob 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@5 -- # export PATH 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@51 -- # : 0 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@55 -- # have_pci_nics=0 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@11 -- # MALLOC_BDEV_SIZE=64 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@12 -- # MALLOC_BLOCK_SIZE=4096 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@14 -- # nvmftestinit 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@474 -- # prepare_net_devs 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@436 -- # local -g is_hw=no 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@438 -- # remove_spdk_ns 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@309 -- # xtrace_disable 00:33:18.339 09:09:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.519 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:33:26.519 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@315 -- # pci_devs=() 00:33:26.519 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@315 -- # local -a pci_devs 00:33:26.519 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@316 -- # pci_net_devs=() 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@317 -- # pci_drivers=() 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@317 -- # local -A pci_drivers 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@319 -- # net_devs=() 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@319 -- # local -ga net_devs 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@320 -- # e810=() 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@320 -- # local -ga e810 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@321 -- # x722=() 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@321 -- # local -ga x722 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@322 -- # mlx=() 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@322 -- # local -ga mlx 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:33:26.520 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:33:26.520 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@416 -- # [[ up == up ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:33:26.520 Found net devices under 0000:4b:00.0: cvl_0_0 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@416 -- # [[ up == up ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:33:26.520 Found net devices under 0000:4b:00.1: cvl_0_1 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@440 -- # is_hw=yes 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:33:26.520 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:33:26.521 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:33:26.521 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.746 ms 00:33:26.521 00:33:26.521 --- 10.0.0.2 ping statistics --- 00:33:26.521 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:26.521 rtt min/avg/max/mdev = 0.746/0.746/0.746/0.000 ms 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:33:26.521 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:33:26.521 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.324 ms 00:33:26.521 00:33:26.521 --- 10.0.0.1 ping statistics --- 00:33:26.521 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:26.521 rtt min/avg/max/mdev = 0.324/0.324/0.324/0.000 ms 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@448 -- # return 0 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@15 -- # nvmfappstart -m 0xE 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@507 -- # nvmfpid=2087364 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@508 -- # waitforlisten 2087364 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xE 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@831 -- # '[' -z 2087364 ']' 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:26.521 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:26.521 09:09:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.521 [2024-11-06 09:09:15.777831] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:33:26.521 [2024-11-06 09:09:15.778957] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:33:26.521 [2024-11-06 09:09:15.779011] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:26.521 [2024-11-06 09:09:15.875790] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:33:26.521 [2024-11-06 09:09:15.916720] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:33:26.521 [2024-11-06 09:09:15.916772] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:33:26.521 [2024-11-06 09:09:15.916781] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:33:26.521 [2024-11-06 09:09:15.916789] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:33:26.521 [2024-11-06 09:09:15.916794] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:33:26.521 [2024-11-06 09:09:15.918423] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:33:26.521 [2024-11-06 09:09:15.918561] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:33:26.521 [2024-11-06 09:09:15.918561] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:33:26.521 [2024-11-06 09:09:15.992709] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:33:26.521 [2024-11-06 09:09:15.992786] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:33:26.521 [2024-11-06 09:09:15.993400] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:33:26.521 [2024-11-06 09:09:15.993708] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@864 -- # return 0 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -a 256 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.521 [2024-11-06 09:09:16.615473] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@20 -- # rpc_cmd bdev_malloc_create 64 4096 -b Malloc0 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:26.521 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.782 Malloc0 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@21 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.782 Delay0 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 Delay0 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.782 [2024-11-06 09:09:16.691414] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:26.782 09:09:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0x1 -t 1 -l warning -q 128 00:33:26.782 [2024-11-06 09:09:16.812390] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:33:29.333 Initializing NVMe Controllers 00:33:29.333 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:33:29.333 controller IO queue size 128 less than required 00:33:29.333 Consider using lower queue depth or small IO size because IO requests may be queued at the NVMe driver. 00:33:29.333 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 0 00:33:29.333 Initialization complete. Launching workers. 00:33:29.333 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 I/O completed: 123, failed: 29191 00:33:29.333 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) abort submitted 29248, failed to submit 66 00:33:29.333 success 29191, unsuccessful 57, failed 0 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- target/abort.sh@38 -- # nvmftestfini 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@514 -- # nvmfcleanup 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@121 -- # sync 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@124 -- # set +e 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@125 -- # for i in {1..20} 00:33:29.333 09:09:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:33:29.333 rmmod nvme_tcp 00:33:29.333 rmmod nvme_fabrics 00:33:29.333 rmmod nvme_keyring 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@128 -- # set -e 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@129 -- # return 0 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@515 -- # '[' -n 2087364 ']' 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@516 -- # killprocess 2087364 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@950 -- # '[' -z 2087364 ']' 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@954 -- # kill -0 2087364 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@955 -- # uname 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2087364 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2087364' 00:33:29.333 killing process with pid 2087364 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@969 -- # kill 2087364 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@974 -- # wait 2087364 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@297 -- # iptr 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@789 -- # iptables-save 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:33:29.333 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@789 -- # iptables-restore 00:33:29.334 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:33:29.334 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@302 -- # remove_spdk_ns 00:33:29.334 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:29.334 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:29.334 09:09:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:31.244 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:33:31.244 00:33:31.244 real 0m13.356s 00:33:31.244 user 0m10.971s 00:33:31.244 sys 0m6.934s 00:33:31.244 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@1126 -- # xtrace_disable 00:33:31.244 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:33:31.244 ************************************ 00:33:31.244 END TEST nvmf_abort 00:33:31.244 ************************************ 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@22 -- # run_test nvmf_ns_hotplug_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp --interrupt-mode 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:33:31.507 ************************************ 00:33:31.507 START TEST nvmf_ns_hotplug_stress 00:33:31.507 ************************************ 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp --interrupt-mode 00:33:31.507 * Looking for test storage... 00:33:31.507 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1689 -- # lcov --version 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@333 -- # local ver1 ver1_l 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@334 -- # local ver2 ver2_l 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # IFS=.-: 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@336 -- # read -ra ver1 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # IFS=.-: 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@337 -- # read -ra ver2 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@338 -- # local 'op=<' 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@340 -- # ver1_l=2 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@341 -- # ver2_l=1 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@344 -- # case "$op" in 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@345 -- # : 1 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v = 0 )) 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # decimal 1 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=1 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 1 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@365 -- # ver1[v]=1 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # decimal 2 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@353 -- # local d=2 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@355 -- # echo 2 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@366 -- # ver2[v]=2 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@368 -- # return 0 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:33:31.507 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:31.507 --rc genhtml_branch_coverage=1 00:33:31.507 --rc genhtml_function_coverage=1 00:33:31.507 --rc genhtml_legend=1 00:33:31.507 --rc geninfo_all_blocks=1 00:33:31.507 --rc geninfo_unexecuted_blocks=1 00:33:31.507 00:33:31.507 ' 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:33:31.507 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:31.507 --rc genhtml_branch_coverage=1 00:33:31.507 --rc genhtml_function_coverage=1 00:33:31.507 --rc genhtml_legend=1 00:33:31.507 --rc geninfo_all_blocks=1 00:33:31.507 --rc geninfo_unexecuted_blocks=1 00:33:31.507 00:33:31.507 ' 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:33:31.507 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:31.507 --rc genhtml_branch_coverage=1 00:33:31.507 --rc genhtml_function_coverage=1 00:33:31.507 --rc genhtml_legend=1 00:33:31.507 --rc geninfo_all_blocks=1 00:33:31.507 --rc geninfo_unexecuted_blocks=1 00:33:31.507 00:33:31.507 ' 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:33:31.507 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:33:31.507 --rc genhtml_branch_coverage=1 00:33:31.507 --rc genhtml_function_coverage=1 00:33:31.507 --rc genhtml_legend=1 00:33:31.507 --rc geninfo_all_blocks=1 00:33:31.507 --rc geninfo_unexecuted_blocks=1 00:33:31.507 00:33:31.507 ' 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # uname -s 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:33:31.507 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@15 -- # shopt -s extglob 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@5 -- # export PATH 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@51 -- # : 0 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@55 -- # have_pci_nics=0 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@22 -- # nvmftestinit 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@474 -- # prepare_net_devs 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@436 -- # local -g is_hw=no 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@438 -- # remove_spdk_ns 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@309 -- # xtrace_disable 00:33:31.770 09:09:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # pci_devs=() 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # local -a pci_devs 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # pci_net_devs=() 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # pci_drivers=() 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # local -A pci_drivers 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # net_devs=() 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@319 -- # local -ga net_devs 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # e810=() 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # local -ga e810 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # x722=() 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # local -ga x722 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # mlx=() 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@322 -- # local -ga mlx 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:33:39.916 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:33:39.916 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:33:39.916 Found net devices under 0000:4b:00.0: cvl_0_0 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:33:39.916 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ up == up ]] 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:33:39.917 Found net devices under 0000:4b:00.1: cvl_0_1 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@440 -- # is_hw=yes 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:33:39.917 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:33:39.917 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.589 ms 00:33:39.917 00:33:39.917 --- 10.0.0.2 ping statistics --- 00:33:39.917 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:39.917 rtt min/avg/max/mdev = 0.589/0.589/0.589/0.000 ms 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:33:39.917 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:33:39.917 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.298 ms 00:33:39.917 00:33:39.917 --- 10.0.0.1 ping statistics --- 00:33:39.917 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:33:39.917 rtt min/avg/max/mdev = 0.298/0.298/0.298/0.000 ms 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@448 -- # return 0 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:33:39.917 09:09:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@23 -- # nvmfappstart -m 0xE 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@724 -- # xtrace_disable 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@507 -- # nvmfpid=2092078 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@508 -- # waitforlisten 2092078 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xE 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@831 -- # '[' -z 2092078 ']' 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@836 -- # local max_retries=100 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:33:39.917 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@840 -- # xtrace_disable 00:33:39.917 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:33:39.917 [2024-11-06 09:09:29.089958] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:33:39.917 [2024-11-06 09:09:29.090941] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:33:39.918 [2024-11-06 09:09:29.090979] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:33:39.918 [2024-11-06 09:09:29.184347] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:33:39.918 [2024-11-06 09:09:29.219024] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:33:39.918 [2024-11-06 09:09:29.219056] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:33:39.918 [2024-11-06 09:09:29.219064] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:33:39.918 [2024-11-06 09:09:29.219071] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:33:39.918 [2024-11-06 09:09:29.219076] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:33:39.918 [2024-11-06 09:09:29.222762] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:33:39.918 [2024-11-06 09:09:29.222897] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:33:39.918 [2024-11-06 09:09:29.222998] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:33:39.918 [2024-11-06 09:09:29.278384] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:33:39.918 [2024-11-06 09:09:29.278538] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:33:39.918 [2024-11-06 09:09:29.279067] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:33:39.918 [2024-11-06 09:09:29.279389] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:33:39.918 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:33:39.918 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@864 -- # return 0 00:33:39.918 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:33:39.918 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@730 -- # xtrace_disable 00:33:39.918 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:33:39.918 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:33:39.918 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@25 -- # null_size=1000 00:33:39.918 09:09:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:33:40.179 [2024-11-06 09:09:30.091844] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:33:40.179 09:09:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:33:40.441 09:09:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:33:40.441 [2024-11-06 09:09:30.476661] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:33:40.441 09:09:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:33:40.702 09:09:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 512 -b Malloc0 00:33:40.962 Malloc0 00:33:40.963 09:09:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:33:40.963 Delay0 00:33:40.963 09:09:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:41.223 09:09:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create NULL1 1000 512 00:33:41.484 NULL1 00:33:41.484 09:09:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:33:41.484 09:09:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@42 -- # PERF_PID=2092500 00:33:41.484 09:09:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 30 -q 128 -w randread -o 512 -Q 1000 00:33:41.484 09:09:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:41.484 09:09:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:41.745 09:09:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:42.006 09:09:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1001 00:33:42.006 09:09:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1001 00:33:42.006 true 00:33:42.267 09:09:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:42.267 09:09:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:42.267 09:09:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:42.527 09:09:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1002 00:33:42.527 09:09:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1002 00:33:42.788 true 00:33:42.788 09:09:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:42.788 09:09:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:42.788 09:09:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:43.048 09:09:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1003 00:33:43.048 09:09:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1003 00:33:43.309 true 00:33:43.309 09:09:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:43.309 09:09:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:43.570 09:09:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:43.570 09:09:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1004 00:33:43.570 09:09:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1004 00:33:43.831 true 00:33:43.831 09:09:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:43.831 09:09:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:44.092 09:09:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:44.092 09:09:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1005 00:33:44.092 09:09:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1005 00:33:44.354 true 00:33:44.354 09:09:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:44.354 09:09:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:44.614 09:09:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:44.875 09:09:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1006 00:33:44.875 09:09:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1006 00:33:44.875 true 00:33:44.875 09:09:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:44.875 09:09:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:45.136 09:09:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:45.397 09:09:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1007 00:33:45.397 09:09:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1007 00:33:45.397 true 00:33:45.397 09:09:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:45.397 09:09:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:45.658 09:09:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:45.919 09:09:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1008 00:33:45.919 09:09:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1008 00:33:45.919 true 00:33:46.180 09:09:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:46.180 09:09:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:46.180 09:09:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:46.441 09:09:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1009 00:33:46.441 09:09:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1009 00:33:46.701 true 00:33:46.701 09:09:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:46.701 09:09:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:46.701 09:09:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:46.961 09:09:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1010 00:33:46.961 09:09:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1010 00:33:47.221 true 00:33:47.221 09:09:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:47.221 09:09:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:47.221 09:09:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:47.481 09:09:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1011 00:33:47.481 09:09:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1011 00:33:47.741 true 00:33:47.741 09:09:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:47.741 09:09:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:48.001 09:09:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:48.001 09:09:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1012 00:33:48.001 09:09:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1012 00:33:48.261 true 00:33:48.261 09:09:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:48.261 09:09:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:48.520 09:09:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:48.520 09:09:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1013 00:33:48.520 09:09:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1013 00:33:48.780 true 00:33:48.780 09:09:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:48.780 09:09:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:49.041 09:09:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:49.301 09:09:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1014 00:33:49.301 09:09:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1014 00:33:49.301 true 00:33:49.301 09:09:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:49.301 09:09:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:49.561 09:09:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:49.821 09:09:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1015 00:33:49.821 09:09:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1015 00:33:49.821 true 00:33:49.821 09:09:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:49.821 09:09:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:50.080 09:09:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:50.340 09:09:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1016 00:33:50.340 09:09:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1016 00:33:50.340 true 00:33:50.600 09:09:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:50.600 09:09:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:50.600 09:09:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:50.861 09:09:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1017 00:33:50.861 09:09:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1017 00:33:51.121 true 00:33:51.121 09:09:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:51.121 09:09:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:51.121 09:09:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:51.382 09:09:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1018 00:33:51.382 09:09:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1018 00:33:51.643 true 00:33:51.643 09:09:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:51.643 09:09:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:51.904 09:09:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:51.904 09:09:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1019 00:33:51.904 09:09:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1019 00:33:52.165 true 00:33:52.165 09:09:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:52.165 09:09:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:52.426 09:09:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:52.426 09:09:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1020 00:33:52.426 09:09:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1020 00:33:52.687 true 00:33:52.687 09:09:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:52.687 09:09:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:52.949 09:09:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:52.949 09:09:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1021 00:33:52.949 09:09:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1021 00:33:53.210 true 00:33:53.210 09:09:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:53.210 09:09:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:53.470 09:09:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:53.732 09:09:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1022 00:33:53.732 09:09:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1022 00:33:53.732 true 00:33:53.732 09:09:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:53.732 09:09:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:53.994 09:09:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:54.255 09:09:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1023 00:33:54.255 09:09:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1023 00:33:54.255 true 00:33:54.255 09:09:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:54.255 09:09:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:54.515 09:09:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:54.775 09:09:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1024 00:33:54.775 09:09:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1024 00:33:54.775 true 00:33:55.035 09:09:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:55.035 09:09:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:55.035 09:09:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:55.296 09:09:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1025 00:33:55.297 09:09:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1025 00:33:55.558 true 00:33:55.558 09:09:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:55.558 09:09:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:55.558 09:09:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:55.818 09:09:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1026 00:33:55.818 09:09:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1026 00:33:56.079 true 00:33:56.079 09:09:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:56.079 09:09:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:56.079 09:09:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:56.339 09:09:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1027 00:33:56.339 09:09:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1027 00:33:56.600 true 00:33:56.600 09:09:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:56.600 09:09:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:56.860 09:09:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:56.860 09:09:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1028 00:33:56.860 09:09:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1028 00:33:57.121 true 00:33:57.121 09:09:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:57.121 09:09:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:57.382 09:09:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:57.382 09:09:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1029 00:33:57.382 09:09:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1029 00:33:57.643 true 00:33:57.643 09:09:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:57.643 09:09:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:57.904 09:09:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:58.165 09:09:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1030 00:33:58.165 09:09:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1030 00:33:58.165 true 00:33:58.165 09:09:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:58.165 09:09:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:58.426 09:09:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:58.687 09:09:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1031 00:33:58.687 09:09:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1031 00:33:58.687 true 00:33:58.687 09:09:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:58.687 09:09:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:58.948 09:09:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:59.209 09:09:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1032 00:33:59.209 09:09:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1032 00:33:59.209 true 00:33:59.209 09:09:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:59.209 09:09:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:59.470 09:09:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:33:59.734 09:09:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1033 00:33:59.734 09:09:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1033 00:33:59.999 true 00:33:59.999 09:09:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:33:59.999 09:09:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:33:59.999 09:09:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:00.258 09:09:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1034 00:34:00.258 09:09:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1034 00:34:00.518 true 00:34:00.518 09:09:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:00.518 09:09:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:00.518 09:09:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:00.777 09:09:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1035 00:34:00.777 09:09:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1035 00:34:01.037 true 00:34:01.037 09:09:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:01.037 09:09:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:01.297 09:09:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:01.297 09:09:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1036 00:34:01.297 09:09:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1036 00:34:01.557 true 00:34:01.557 09:09:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:01.557 09:09:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:01.816 09:09:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:01.816 09:09:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1037 00:34:01.816 09:09:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1037 00:34:02.076 true 00:34:02.076 09:09:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:02.076 09:09:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:02.336 09:09:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:02.336 09:09:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1038 00:34:02.596 09:09:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1038 00:34:02.596 true 00:34:02.596 09:09:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:02.596 09:09:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:02.857 09:09:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:03.117 09:09:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1039 00:34:03.117 09:09:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1039 00:34:03.117 true 00:34:03.117 09:09:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:03.117 09:09:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:03.378 09:09:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:03.637 09:09:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1040 00:34:03.637 09:09:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1040 00:34:03.637 true 00:34:03.637 09:09:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:03.637 09:09:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:03.898 09:09:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:04.158 09:09:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1041 00:34:04.158 09:09:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1041 00:34:04.158 true 00:34:04.418 09:09:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:04.418 09:09:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:04.418 09:09:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:04.678 09:09:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1042 00:34:04.678 09:09:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1042 00:34:04.938 true 00:34:04.938 09:09:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:04.938 09:09:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:04.938 09:09:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:05.198 09:09:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1043 00:34:05.198 09:09:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1043 00:34:05.458 true 00:34:05.458 09:09:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:05.458 09:09:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:05.717 09:09:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:05.717 09:09:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1044 00:34:05.717 09:09:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1044 00:34:05.977 true 00:34:05.977 09:09:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:05.977 09:09:55 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:06.237 09:09:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:06.237 09:09:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1045 00:34:06.237 09:09:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1045 00:34:06.498 true 00:34:06.498 09:09:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:06.498 09:09:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:06.759 09:09:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:07.019 09:09:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1046 00:34:07.019 09:09:56 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1046 00:34:07.019 true 00:34:07.019 09:09:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:07.019 09:09:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:07.280 09:09:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:07.540 09:09:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1047 00:34:07.540 09:09:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1047 00:34:07.540 true 00:34:07.540 09:09:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:07.540 09:09:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:07.799 09:09:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:08.059 09:09:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1048 00:34:08.059 09:09:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1048 00:34:08.320 true 00:34:08.320 09:09:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:08.320 09:09:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:08.320 09:09:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:08.581 09:09:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1049 00:34:08.581 09:09:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1049 00:34:08.841 true 00:34:08.841 09:09:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:08.841 09:09:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:08.841 09:09:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:09.102 09:09:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1050 00:34:09.102 09:09:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1050 00:34:09.362 true 00:34:09.362 09:09:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:09.362 09:09:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:09.621 09:09:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:09.622 09:09:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1051 00:34:09.622 09:09:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1051 00:34:09.882 true 00:34:09.882 09:09:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:09.882 09:09:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:10.142 09:10:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:10.142 09:10:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1052 00:34:10.142 09:10:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1052 00:34:10.403 true 00:34:10.403 09:10:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:10.403 09:10:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:10.663 09:10:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:10.924 09:10:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1053 00:34:10.924 09:10:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1053 00:34:10.924 true 00:34:10.924 09:10:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:10.924 09:10:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:11.185 09:10:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:11.445 09:10:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1054 00:34:11.445 09:10:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1054 00:34:11.445 true 00:34:11.445 09:10:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:11.445 09:10:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:11.761 09:10:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:12.082 09:10:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1055 00:34:12.082 09:10:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1055 00:34:12.082 Initializing NVMe Controllers 00:34:12.082 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:34:12.082 Controller IO queue size 128, less than required. 00:34:12.082 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:34:12.082 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:34:12.082 Initialization complete. Launching workers. 00:34:12.082 ======================================================== 00:34:12.082 Latency(us) 00:34:12.082 Device Information : IOPS MiB/s Average min max 00:34:12.082 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 30236.31 14.76 4233.17 1500.87 11159.47 00:34:12.082 ======================================================== 00:34:12.082 Total : 30236.31 14.76 4233.17 1500.87 11159.47 00:34:12.082 00:34:12.082 true 00:34:12.082 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 2092500 00:34:12.082 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh: line 44: kill: (2092500) - No such process 00:34:12.082 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@53 -- # wait 2092500 00:34:12.082 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:12.362 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:12.362 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # nthreads=8 00:34:12.362 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # pids=() 00:34:12.362 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i = 0 )) 00:34:12.362 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:34:12.362 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null0 100 4096 00:34:12.622 null0 00:34:12.622 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:34:12.622 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:34:12.622 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null1 100 4096 00:34:12.882 null1 00:34:12.882 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:34:12.882 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:34:12.882 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null2 100 4096 00:34:12.882 null2 00:34:12.882 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:34:12.882 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:34:12.882 09:10:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null3 100 4096 00:34:13.144 null3 00:34:13.144 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:34:13.144 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:34:13.144 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null4 100 4096 00:34:13.405 null4 00:34:13.405 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:34:13.405 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:34:13.405 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null5 100 4096 00:34:13.405 null5 00:34:13.405 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:34:13.405 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:34:13.405 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null6 100 4096 00:34:13.666 null6 00:34:13.666 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:34:13.666 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:34:13.666 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null7 100 4096 00:34:13.929 null7 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i = 0 )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 1 null0 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=1 bdev=null0 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 2 null1 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=2 bdev=null1 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 3 null2 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=3 bdev=null2 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 4 null3 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=4 bdev=null3 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 5 null4 00:34:13.929 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=5 bdev=null4 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 6 null5 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=6 bdev=null5 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 7 null6 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=7 bdev=null6 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@66 -- # wait 2098854 2098856 2098859 2098862 2098865 2098868 2098870 2098872 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 8 null7 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=8 bdev=null7 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:13.930 09:10:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.191 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:34:14.451 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.712 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.713 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:34:14.713 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.713 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.713 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:34:14.713 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.713 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.713 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:34:14.713 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:34:14.713 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:14.713 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.973 09:10:04 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:34:14.973 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.973 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.973 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.973 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:14.974 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.235 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:34:15.496 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:15.757 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:34:16.019 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:16.019 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:34:16.019 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:34:16.019 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:34:16.019 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.019 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.019 09:10:05 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.019 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:34:16.280 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:34:16.541 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:34:16.803 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:34:17.064 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.064 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.064 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:34:17.064 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.064 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.064 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:34:17.064 09:10:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:17.064 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:34:17.064 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.064 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.064 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:34:17.064 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:34:17.064 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:34:17.064 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:34:17.064 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:34:17.064 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:17.325 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.325 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.325 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:34:17.325 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.325 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.325 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:34:17.325 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.325 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.325 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:34:17.326 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.587 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@70 -- # nvmftestfini 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@514 -- # nvmfcleanup 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@121 -- # sync 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@124 -- # set +e 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@125 -- # for i in {1..20} 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:34:17.849 rmmod nvme_tcp 00:34:17.849 rmmod nvme_fabrics 00:34:17.849 rmmod nvme_keyring 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@128 -- # set -e 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@129 -- # return 0 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@515 -- # '[' -n 2092078 ']' 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@516 -- # killprocess 2092078 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@950 -- # '[' -z 2092078 ']' 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@954 -- # kill -0 2092078 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@955 -- # uname 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2092078 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2092078' 00:34:17.849 killing process with pid 2092078 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@969 -- # kill 2092078 00:34:17.849 09:10:07 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@974 -- # wait 2092078 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@297 -- # iptr 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # iptables-save 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@789 -- # iptables-restore 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@302 -- # remove_spdk_ns 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:18.111 09:10:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:20.025 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:34:20.286 00:34:20.286 real 0m48.733s 00:34:20.286 user 3m1.647s 00:34:20.286 sys 0m22.707s 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:34:20.286 ************************************ 00:34:20.286 END TEST nvmf_ns_hotplug_stress 00:34:20.286 ************************************ 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@23 -- # run_test nvmf_delete_subsystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp --interrupt-mode 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:34:20.286 ************************************ 00:34:20.286 START TEST nvmf_delete_subsystem 00:34:20.286 ************************************ 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp --interrupt-mode 00:34:20.286 * Looking for test storage... 00:34:20.286 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1689 -- # lcov --version 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@333 -- # local ver1 ver1_l 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@334 -- # local ver2 ver2_l 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@336 -- # IFS=.-: 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@336 -- # read -ra ver1 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@337 -- # IFS=.-: 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@337 -- # read -ra ver2 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@338 -- # local 'op=<' 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@340 -- # ver1_l=2 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@341 -- # ver2_l=1 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@344 -- # case "$op" in 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@345 -- # : 1 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v = 0 )) 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:34:20.286 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@365 -- # decimal 1 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=1 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 1 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@365 -- # ver1[v]=1 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@366 -- # decimal 2 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@353 -- # local d=2 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@355 -- # echo 2 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@366 -- # ver2[v]=2 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@368 -- # return 0 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:34:20.547 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:20.547 --rc genhtml_branch_coverage=1 00:34:20.547 --rc genhtml_function_coverage=1 00:34:20.547 --rc genhtml_legend=1 00:34:20.547 --rc geninfo_all_blocks=1 00:34:20.547 --rc geninfo_unexecuted_blocks=1 00:34:20.547 00:34:20.547 ' 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:34:20.547 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:20.547 --rc genhtml_branch_coverage=1 00:34:20.547 --rc genhtml_function_coverage=1 00:34:20.547 --rc genhtml_legend=1 00:34:20.547 --rc geninfo_all_blocks=1 00:34:20.547 --rc geninfo_unexecuted_blocks=1 00:34:20.547 00:34:20.547 ' 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:34:20.547 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:20.547 --rc genhtml_branch_coverage=1 00:34:20.547 --rc genhtml_function_coverage=1 00:34:20.547 --rc genhtml_legend=1 00:34:20.547 --rc geninfo_all_blocks=1 00:34:20.547 --rc geninfo_unexecuted_blocks=1 00:34:20.547 00:34:20.547 ' 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:34:20.547 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:20.547 --rc genhtml_branch_coverage=1 00:34:20.547 --rc genhtml_function_coverage=1 00:34:20.547 --rc genhtml_legend=1 00:34:20.547 --rc geninfo_all_blocks=1 00:34:20.547 --rc geninfo_unexecuted_blocks=1 00:34:20.547 00:34:20.547 ' 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # uname -s 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:20.547 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@15 -- # shopt -s extglob 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@5 -- # export PATH 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@51 -- # : 0 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@55 -- # have_pci_nics=0 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@12 -- # nvmftestinit 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@474 -- # prepare_net_devs 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@436 -- # local -g is_hw=no 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@438 -- # remove_spdk_ns 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@309 -- # xtrace_disable 00:34:20.548 09:10:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # pci_devs=() 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # local -a pci_devs 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # pci_net_devs=() 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # pci_drivers=() 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # local -A pci_drivers 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # net_devs=() 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@319 -- # local -ga net_devs 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # e810=() 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # local -ga e810 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # x722=() 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # local -ga x722 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # mlx=() 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@322 -- # local -ga mlx 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:34:27.130 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:27.130 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:34:27.131 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:34:27.131 Found net devices under 0000:4b:00.0: cvl_0_0 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ up == up ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:34:27.131 Found net devices under 0000:4b:00.1: cvl_0_1 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@440 -- # is_hw=yes 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:34:27.131 09:10:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:34:27.131 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:34:27.131 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:34:27.131 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:34:27.131 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:34:27.131 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:34:27.131 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:34:27.131 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:34:27.131 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:34:27.131 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:34:27.131 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.661 ms 00:34:27.131 00:34:27.131 --- 10.0.0.2 ping statistics --- 00:34:27.131 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:27.131 rtt min/avg/max/mdev = 0.661/0.661/0.661/0.000 ms 00:34:27.131 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:34:27.131 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:34:27.131 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.262 ms 00:34:27.131 00:34:27.131 --- 10.0.0.1 ping statistics --- 00:34:27.131 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:27.131 rtt min/avg/max/mdev = 0.262/0.262/0.262/0.000 ms 00:34:27.392 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:34:27.392 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@448 -- # return 0 00:34:27.392 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@13 -- # nvmfappstart -m 0x3 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@724 -- # xtrace_disable 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@507 -- # nvmfpid=2103779 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@508 -- # waitforlisten 2103779 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x3 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@831 -- # '[' -z 2103779 ']' 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@836 -- # local max_retries=100 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:27.393 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@840 -- # xtrace_disable 00:34:27.393 09:10:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:27.393 [2024-11-06 09:10:17.346596] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:34:27.393 [2024-11-06 09:10:17.347568] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:34:27.393 [2024-11-06 09:10:17.347606] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:27.393 [2024-11-06 09:10:17.424968] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:34:27.393 [2024-11-06 09:10:17.459468] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:34:27.393 [2024-11-06 09:10:17.459503] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:34:27.393 [2024-11-06 09:10:17.459512] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:34:27.393 [2024-11-06 09:10:17.459523] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:34:27.393 [2024-11-06 09:10:17.459528] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:34:27.393 [2024-11-06 09:10:17.460671] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:34:27.393 [2024-11-06 09:10:17.460673] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:27.653 [2024-11-06 09:10:17.516586] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:34:27.653 [2024-11-06 09:10:17.517048] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:34:27.653 [2024-11-06 09:10:17.517402] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@864 -- # return 0 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:28.250 [2024-11-06 09:10:18.201494] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:28.250 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:28.251 [2024-11-06 09:10:18.229976] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:28.251 NULL1 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@23 -- # rpc_cmd bdev_delay_create -b NULL1 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:28.251 Delay0 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@28 -- # perf_pid=2104082 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@30 -- # sleep 2 00:34:28.251 09:10:18 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 5 -q 128 -w randrw -M 70 -o 512 -P 4 00:34:28.251 [2024-11-06 09:10:18.328569] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:34:30.166 09:10:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@32 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:34:30.166 09:10:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:30.166 09:10:20 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 starting I/O failed: -6 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 [2024-11-06 09:10:20.456438] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1cdc2c0 is same with the state(6) to be set 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.426 Write completed with error (sct=0, sc=8) 00:34:30.426 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 starting I/O failed: -6 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 starting I/O failed: -6 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 starting I/O failed: -6 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 starting I/O failed: -6 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 starting I/O failed: -6 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 starting I/O failed: -6 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 starting I/O failed: -6 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 starting I/O failed: -6 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 starting I/O failed: -6 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 [2024-11-06 09:10:20.459494] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7ff89c00d450 is same with the state(6) to be set 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Write completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:30.427 Read completed with error (sct=0, sc=8) 00:34:31.370 [2024-11-06 09:10:21.426936] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1cdd9a0 is same with the state(6) to be set 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 [2024-11-06 09:10:21.460007] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1cdc680 is same with the state(6) to be set 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 [2024-11-06 09:10:21.460665] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1cdc860 is same with the state(6) to be set 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 [2024-11-06 09:10:21.461161] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7ff89c00cfe0 is same with the state(6) to be set 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 Read completed with error (sct=0, sc=8) 00:34:31.370 Write completed with error (sct=0, sc=8) 00:34:31.370 [2024-11-06 09:10:21.461248] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7ff89c00d780 is same with the state(6) to be set 00:34:31.370 Initializing NVMe Controllers 00:34:31.370 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:34:31.370 Controller IO queue size 128, less than required. 00:34:31.370 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:34:31.370 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:34:31.370 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:34:31.370 Initialization complete. Launching workers. 00:34:31.370 ======================================================== 00:34:31.370 Latency(us) 00:34:31.370 Device Information : IOPS MiB/s Average min max 00:34:31.370 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 173.68 0.08 886199.34 242.69 1008485.10 00:34:31.370 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 153.78 0.08 982158.96 263.28 2001852.75 00:34:31.370 ======================================================== 00:34:31.370 Total : 327.46 0.16 931262.45 242.69 2001852.75 00:34:31.370 00:34:31.370 [2024-11-06 09:10:21.462063] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1cdd9a0 (9): Bad file descriptor 00:34:31.370 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:34:31.370 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:31.371 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@34 -- # delay=0 00:34:31.371 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 2104082 00:34:31.371 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@36 -- # sleep 0.5 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@38 -- # (( delay++ > 30 )) 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 2104082 00:34:31.942 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 35: kill: (2104082) - No such process 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@45 -- # NOT wait 2104082 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@650 -- # local es=0 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@652 -- # valid_exec_arg wait 2104082 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@638 -- # local arg=wait 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # type -t wait 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@653 -- # wait 2104082 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@653 -- # es=1 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:31.942 [2024-11-06 09:10:21.993956] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@50 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:31.942 09:10:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:31.942 09:10:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:31.942 09:10:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@54 -- # perf_pid=2104802 00:34:31.942 09:10:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@56 -- # delay=0 00:34:31.942 09:10:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 3 -q 128 -w randrw -M 70 -o 512 -P 4 00:34:31.942 09:10:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 2104802 00:34:31.942 09:10:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:34:32.203 [2024-11-06 09:10:22.064453] subsystem.c:1637:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:34:32.464 09:10:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:34:32.464 09:10:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 2104802 00:34:32.464 09:10:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:34:33.035 09:10:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:34:33.035 09:10:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 2104802 00:34:33.035 09:10:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:34:33.608 09:10:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:34:33.608 09:10:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 2104802 00:34:33.608 09:10:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:34:34.179 09:10:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:34:34.179 09:10:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 2104802 00:34:34.179 09:10:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:34:34.439 09:10:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:34:34.439 09:10:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 2104802 00:34:34.439 09:10:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:34:35.011 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:34:35.011 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 2104802 00:34:35.011 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:34:35.272 Initializing NVMe Controllers 00:34:35.272 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:34:35.272 Controller IO queue size 128, less than required. 00:34:35.272 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:34:35.272 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:34:35.272 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:34:35.272 Initialization complete. Launching workers. 00:34:35.272 ======================================================== 00:34:35.272 Latency(us) 00:34:35.272 Device Information : IOPS MiB/s Average min max 00:34:35.272 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 128.00 0.06 1002623.23 1000147.86 1041397.36 00:34:35.272 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 128.00 0.06 1004272.72 1000329.20 1010284.16 00:34:35.272 ======================================================== 00:34:35.272 Total : 256.00 0.12 1003447.98 1000147.86 1041397.36 00:34:35.272 00:34:35.533 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:34:35.533 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 2104802 00:34:35.533 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 57: kill: (2104802) - No such process 00:34:35.533 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@67 -- # wait 2104802 00:34:35.533 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:34:35.533 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- target/delete_subsystem.sh@71 -- # nvmftestfini 00:34:35.533 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@514 -- # nvmfcleanup 00:34:35.533 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@121 -- # sync 00:34:35.533 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:34:35.533 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@124 -- # set +e 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@125 -- # for i in {1..20} 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:34:35.534 rmmod nvme_tcp 00:34:35.534 rmmod nvme_fabrics 00:34:35.534 rmmod nvme_keyring 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@128 -- # set -e 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@129 -- # return 0 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@515 -- # '[' -n 2103779 ']' 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@516 -- # killprocess 2103779 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@950 -- # '[' -z 2103779 ']' 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@954 -- # kill -0 2103779 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@955 -- # uname 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:34:35.534 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2103779 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2103779' 00:34:35.795 killing process with pid 2103779 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@969 -- # kill 2103779 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@974 -- # wait 2103779 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@297 -- # iptr 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # iptables-save 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@789 -- # iptables-restore 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@302 -- # remove_spdk_ns 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:35.795 09:10:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:38.343 09:10:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:34:38.343 00:34:38.343 real 0m17.668s 00:34:38.343 user 0m26.235s 00:34:38.343 sys 0m7.040s 00:34:38.343 09:10:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:38.343 09:10:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:34:38.343 ************************************ 00:34:38.343 END TEST nvmf_delete_subsystem 00:34:38.343 ************************************ 00:34:38.343 09:10:27 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@26 -- # run_test nvmf_host_management /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp --interrupt-mode 00:34:38.343 09:10:27 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:34:38.343 09:10:27 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:38.343 09:10:27 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:34:38.343 ************************************ 00:34:38.343 START TEST nvmf_host_management 00:34:38.343 ************************************ 00:34:38.343 09:10:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp --interrupt-mode 00:34:38.343 * Looking for test storage... 00:34:38.343 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1689 -- # lcov --version 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@333 -- # local ver1 ver1_l 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@334 -- # local ver2 ver2_l 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@336 -- # IFS=.-: 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@336 -- # read -ra ver1 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@337 -- # IFS=.-: 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@337 -- # read -ra ver2 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@338 -- # local 'op=<' 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@340 -- # ver1_l=2 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@341 -- # ver2_l=1 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@344 -- # case "$op" in 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@345 -- # : 1 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@364 -- # (( v = 0 )) 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@365 -- # decimal 1 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@353 -- # local d=1 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@355 -- # echo 1 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@365 -- # ver1[v]=1 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@366 -- # decimal 2 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@353 -- # local d=2 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@355 -- # echo 2 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@366 -- # ver2[v]=2 00:34:38.343 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@368 -- # return 0 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:34:38.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:38.344 --rc genhtml_branch_coverage=1 00:34:38.344 --rc genhtml_function_coverage=1 00:34:38.344 --rc genhtml_legend=1 00:34:38.344 --rc geninfo_all_blocks=1 00:34:38.344 --rc geninfo_unexecuted_blocks=1 00:34:38.344 00:34:38.344 ' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:34:38.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:38.344 --rc genhtml_branch_coverage=1 00:34:38.344 --rc genhtml_function_coverage=1 00:34:38.344 --rc genhtml_legend=1 00:34:38.344 --rc geninfo_all_blocks=1 00:34:38.344 --rc geninfo_unexecuted_blocks=1 00:34:38.344 00:34:38.344 ' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:34:38.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:38.344 --rc genhtml_branch_coverage=1 00:34:38.344 --rc genhtml_function_coverage=1 00:34:38.344 --rc genhtml_legend=1 00:34:38.344 --rc geninfo_all_blocks=1 00:34:38.344 --rc geninfo_unexecuted_blocks=1 00:34:38.344 00:34:38.344 ' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:34:38.344 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:38.344 --rc genhtml_branch_coverage=1 00:34:38.344 --rc genhtml_function_coverage=1 00:34:38.344 --rc genhtml_legend=1 00:34:38.344 --rc geninfo_all_blocks=1 00:34:38.344 --rc geninfo_unexecuted_blocks=1 00:34:38.344 00:34:38.344 ' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@7 -- # uname -s 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@15 -- # shopt -s extglob 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@5 -- # export PATH 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@51 -- # : 0 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@55 -- # have_pci_nics=0 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@11 -- # MALLOC_BDEV_SIZE=64 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@105 -- # nvmftestinit 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@474 -- # prepare_net_devs 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@436 -- # local -g is_hw=no 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@438 -- # remove_spdk_ns 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@309 -- # xtrace_disable 00:34:38.344 09:10:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:46.490 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@315 -- # pci_devs=() 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@315 -- # local -a pci_devs 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@316 -- # pci_net_devs=() 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@317 -- # pci_drivers=() 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@317 -- # local -A pci_drivers 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@319 -- # net_devs=() 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@319 -- # local -ga net_devs 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@320 -- # e810=() 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@320 -- # local -ga e810 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@321 -- # x722=() 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@321 -- # local -ga x722 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@322 -- # mlx=() 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@322 -- # local -ga mlx 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:34:46.491 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:34:46.491 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@416 -- # [[ up == up ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:34:46.491 Found net devices under 0000:4b:00.0: cvl_0_0 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@416 -- # [[ up == up ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:34:46.491 Found net devices under 0000:4b:00.1: cvl_0_1 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@440 -- # is_hw=yes 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:34:46.491 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:34:46.492 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:34:46.492 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.504 ms 00:34:46.492 00:34:46.492 --- 10.0.0.2 ping statistics --- 00:34:46.492 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:46.492 rtt min/avg/max/mdev = 0.504/0.504/0.504/0.000 ms 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:34:46.492 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:34:46.492 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.259 ms 00:34:46.492 00:34:46.492 --- 10.0.0.1 ping statistics --- 00:34:46.492 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:34:46.492 rtt min/avg/max/mdev = 0.259/0.259/0.259/0.000 ms 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@448 -- # return 0 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@107 -- # nvmf_host_management 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@69 -- # starttarget 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@16 -- # nvmfappstart -m 0x1E 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@724 -- # xtrace_disable 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@507 -- # nvmfpid=2109484 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@508 -- # waitforlisten 2109484 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1E 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@831 -- # '[' -z 2109484 ']' 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@836 -- # local max_retries=100 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:34:46.492 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@840 -- # xtrace_disable 00:34:46.492 09:10:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:46.492 [2024-11-06 09:10:35.521114] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:34:46.492 [2024-11-06 09:10:35.522133] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:34:46.492 [2024-11-06 09:10:35.522174] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:34:46.492 [2024-11-06 09:10:35.619890] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:34:46.492 [2024-11-06 09:10:35.672359] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:34:46.492 [2024-11-06 09:10:35.672418] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:34:46.492 [2024-11-06 09:10:35.672427] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:34:46.492 [2024-11-06 09:10:35.672434] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:34:46.492 [2024-11-06 09:10:35.672440] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:34:46.492 [2024-11-06 09:10:35.674460] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:34:46.492 [2024-11-06 09:10:35.674630] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:34:46.492 [2024-11-06 09:10:35.674813] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:34:46.492 [2024-11-06 09:10:35.674847] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:34:46.492 [2024-11-06 09:10:35.754960] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:34:46.492 [2024-11-06 09:10:35.755606] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:34:46.492 [2024-11-06 09:10:35.756610] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:34:46.492 [2024-11-06 09:10:35.756627] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:34:46.492 [2024-11-06 09:10:35.756784] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@864 -- # return 0 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:46.492 [2024-11-06 09:10:36.388181] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@20 -- # timing_enter create_subsystem 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@724 -- # xtrace_disable 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@22 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@23 -- # cat 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@30 -- # rpc_cmd 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:46.492 Malloc0 00:34:46.492 [2024-11-06 09:10:36.492519] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@31 -- # timing_exit create_subsystems 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@730 -- # xtrace_disable 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@73 -- # perfpid=2109850 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@74 -- # waitforlisten 2109850 /var/tmp/bdevperf.sock 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@831 -- # '[' -z 2109850 ']' 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@836 -- # local max_retries=100 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:34:46.492 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:34:46.492 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@72 -- # gen_nvmf_target_json 0 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@840 -- # xtrace_disable 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@558 -- # config=() 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@558 -- # local subsystem config 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:34:46.493 { 00:34:46.493 "params": { 00:34:46.493 "name": "Nvme$subsystem", 00:34:46.493 "trtype": "$TEST_TRANSPORT", 00:34:46.493 "traddr": "$NVMF_FIRST_TARGET_IP", 00:34:46.493 "adrfam": "ipv4", 00:34:46.493 "trsvcid": "$NVMF_PORT", 00:34:46.493 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:34:46.493 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:34:46.493 "hdgst": ${hdgst:-false}, 00:34:46.493 "ddgst": ${ddgst:-false} 00:34:46.493 }, 00:34:46.493 "method": "bdev_nvme_attach_controller" 00:34:46.493 } 00:34:46.493 EOF 00:34:46.493 )") 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@580 -- # cat 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # jq . 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@583 -- # IFS=, 00:34:46.493 09:10:36 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:34:46.493 "params": { 00:34:46.493 "name": "Nvme0", 00:34:46.493 "trtype": "tcp", 00:34:46.493 "traddr": "10.0.0.2", 00:34:46.493 "adrfam": "ipv4", 00:34:46.493 "trsvcid": "4420", 00:34:46.493 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:46.493 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:34:46.493 "hdgst": false, 00:34:46.493 "ddgst": false 00:34:46.493 }, 00:34:46.493 "method": "bdev_nvme_attach_controller" 00:34:46.493 }' 00:34:46.753 [2024-11-06 09:10:36.610937] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:34:46.753 [2024-11-06 09:10:36.611004] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2109850 ] 00:34:46.753 [2024-11-06 09:10:36.682152] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:46.753 [2024-11-06 09:10:36.718352] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:46.753 Running I/O for 10 seconds... 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@864 -- # return 0 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@75 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@78 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@80 -- # waitforio /var/tmp/bdevperf.sock Nvme0n1 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@45 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@49 -- # '[' -z Nvme0n1 ']' 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@52 -- # local ret=1 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@53 -- # local i 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@54 -- # (( i = 10 )) 00:34:47.325 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=938 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@58 -- # '[' 938 -ge 100 ']' 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@59 -- # ret=0 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@60 -- # break 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@64 -- # return 0 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@84 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:47.587 [2024-11-06 09:10:37.484309] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1eb1f20 is same with the state(6) to be set 00:34:47.587 [2024-11-06 09:10:37.484352] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1eb1f20 is same with the state(6) to be set 00:34:47.587 [2024-11-06 09:10:37.484360] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1eb1f20 is same with the state(6) to be set 00:34:47.587 [2024-11-06 09:10:37.484367] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1eb1f20 is same with the state(6) to be set 00:34:47.587 [2024-11-06 09:10:37.484374] tcp.c:1773:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1eb1f20 is same with the state(6) to be set 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@85 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@561 -- # xtrace_disable 00:34:47.587 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:47.587 [2024-11-06 09:10:37.497774] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:34:47.587 [2024-11-06 09:10:37.497810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.587 [2024-11-06 09:10:37.497826] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:34:47.587 [2024-11-06 09:10:37.497835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.587 [2024-11-06 09:10:37.497843] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:34:47.588 [2024-11-06 09:10:37.497851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.497859] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:34:47.588 [2024-11-06 09:10:37.497866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.497874] nvme_tcp.c: 326:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1f6a000 is same with the state(6) to be set 00:34:47.588 [2024-11-06 09:10:37.497949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:130304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.497958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.497972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:130432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.497980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.497990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:130560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.497997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:130688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:130816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:130944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498066] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:1024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:1152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:1280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:1408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498264] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:1536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:1664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:1792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:2048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:2176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:2304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:2432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:2560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498408] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:2688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:2816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:2944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498459] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:3072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498485] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:3200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:3328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:3456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:3584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:3712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:3840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.588 [2024-11-06 09:10:37.498583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.588 [2024-11-06 09:10:37.498592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:3968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:4096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:4224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:4352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:4480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:4608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:4736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:4864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:4992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:5120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:5248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:5376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:5504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:5632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:5760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:5888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:6016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:6144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:6272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498923] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:6400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:6528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:6656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.498984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:6784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.498992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.499001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:6912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.499009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.499018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:7040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.499026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.499035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:7168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.499042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.499052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:7296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:34:47.589 [2024-11-06 09:10:37.499060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:34:47.589 [2024-11-06 09:10:37.500291] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 1] resetting controller 00:34:47.589 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:34:47.589 09:10:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@87 -- # sleep 1 00:34:47.589 task offset: 130304 on job bdev=Nvme0n1 fails 00:34:47.589 00:34:47.589 Latency(us) 00:34:47.589 [2024-11-06T08:10:37.702Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:47.589 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:34:47.589 Job: Nvme0n1 ended in about 0.64 seconds with error 00:34:47.589 Verification LBA range: start 0x0 length 0x400 00:34:47.589 Nvme0n1 : 0.64 1595.17 99.70 100.29 0.00 36902.69 1672.53 32331.09 00:34:47.589 [2024-11-06T08:10:37.702Z] =================================================================================================================== 00:34:47.589 [2024-11-06T08:10:37.702Z] Total : 1595.17 99.70 100.29 0.00 36902.69 1672.53 32331.09 00:34:47.589 [2024-11-06 09:10:37.502280] app.c:1064:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:34:47.589 [2024-11-06 09:10:37.502300] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1f6a000 (9): Bad file descriptor 00:34:47.589 [2024-11-06 09:10:37.507641] bdev_nvme.c:2236:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode0, 2] Resetting controller successful. 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@91 -- # kill -9 2109850 00:34:48.532 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh: line 91: kill: (2109850) - No such process 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@91 -- # true 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@97 -- # rm -f /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 /var/tmp/spdk_cpu_lock_003 /var/tmp/spdk_cpu_lock_004 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@100 -- # gen_nvmf_target_json 0 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@558 -- # config=() 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@558 -- # local subsystem config 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:34:48.532 { 00:34:48.532 "params": { 00:34:48.532 "name": "Nvme$subsystem", 00:34:48.532 "trtype": "$TEST_TRANSPORT", 00:34:48.532 "traddr": "$NVMF_FIRST_TARGET_IP", 00:34:48.532 "adrfam": "ipv4", 00:34:48.532 "trsvcid": "$NVMF_PORT", 00:34:48.532 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:34:48.532 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:34:48.532 "hdgst": ${hdgst:-false}, 00:34:48.532 "ddgst": ${ddgst:-false} 00:34:48.532 }, 00:34:48.532 "method": "bdev_nvme_attach_controller" 00:34:48.532 } 00:34:48.532 EOF 00:34:48.532 )") 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@580 -- # cat 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@582 -- # jq . 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@583 -- # IFS=, 00:34:48.532 09:10:38 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:34:48.532 "params": { 00:34:48.532 "name": "Nvme0", 00:34:48.532 "trtype": "tcp", 00:34:48.532 "traddr": "10.0.0.2", 00:34:48.532 "adrfam": "ipv4", 00:34:48.532 "trsvcid": "4420", 00:34:48.532 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:34:48.532 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:34:48.532 "hdgst": false, 00:34:48.532 "ddgst": false 00:34:48.532 }, 00:34:48.532 "method": "bdev_nvme_attach_controller" 00:34:48.532 }' 00:34:48.532 [2024-11-06 09:10:38.568208] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:34:48.532 [2024-11-06 09:10:38.568266] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2110201 ] 00:34:48.532 [2024-11-06 09:10:38.638852] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:34:48.793 [2024-11-06 09:10:38.674351] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:34:49.054 Running I/O for 1 seconds... 00:34:49.997 1840.00 IOPS, 115.00 MiB/s 00:34:49.997 Latency(us) 00:34:49.997 [2024-11-06T08:10:40.110Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:34:49.997 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:34:49.997 Verification LBA range: start 0x0 length 0x400 00:34:49.997 Nvme0n1 : 1.01 1884.89 117.81 0.00 0.00 33220.88 1925.12 38010.88 00:34:49.997 [2024-11-06T08:10:40.110Z] =================================================================================================================== 00:34:49.998 [2024-11-06T08:10:40.111Z] Total : 1884.89 117.81 0.00 0.00 33220.88 1925.12 38010.88 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@102 -- # stoptarget 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@36 -- # rm -f ./local-job0-0-verify.state 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@37 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@38 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@40 -- # nvmftestfini 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@514 -- # nvmfcleanup 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@121 -- # sync 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@124 -- # set +e 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@125 -- # for i in {1..20} 00:34:49.998 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:34:49.998 rmmod nvme_tcp 00:34:49.998 rmmod nvme_fabrics 00:34:50.258 rmmod nvme_keyring 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@128 -- # set -e 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@129 -- # return 0 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@515 -- # '[' -n 2109484 ']' 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@516 -- # killprocess 2109484 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@950 -- # '[' -z 2109484 ']' 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@954 -- # kill -0 2109484 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@955 -- # uname 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2109484 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2109484' 00:34:50.258 killing process with pid 2109484 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@969 -- # kill 2109484 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@974 -- # wait 2109484 00:34:50.258 [2024-11-06 09:10:40.321234] app.c: 721:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 1, errno: 2 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@297 -- # iptr 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@789 -- # iptables-save 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@789 -- # iptables-restore 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@302 -- # remove_spdk_ns 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:50.258 09:10:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- target/host_management.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:34:52.804 00:34:52.804 real 0m14.476s 00:34:52.804 user 0m19.302s 00:34:52.804 sys 0m7.284s 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@1126 -- # xtrace_disable 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:34:52.804 ************************************ 00:34:52.804 END TEST nvmf_host_management 00:34:52.804 ************************************ 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@27 -- # run_test nvmf_lvol /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp --interrupt-mode 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:34:52.804 ************************************ 00:34:52.804 START TEST nvmf_lvol 00:34:52.804 ************************************ 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp --interrupt-mode 00:34:52.804 * Looking for test storage... 00:34:52.804 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1689 -- # lcov --version 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@333 -- # local ver1 ver1_l 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@334 -- # local ver2 ver2_l 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@336 -- # IFS=.-: 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@336 -- # read -ra ver1 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@337 -- # IFS=.-: 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@337 -- # read -ra ver2 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@338 -- # local 'op=<' 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@340 -- # ver1_l=2 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@341 -- # ver2_l=1 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@344 -- # case "$op" in 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@345 -- # : 1 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@364 -- # (( v = 0 )) 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@365 -- # decimal 1 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@353 -- # local d=1 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@355 -- # echo 1 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@365 -- # ver1[v]=1 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@366 -- # decimal 2 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@353 -- # local d=2 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@355 -- # echo 2 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@366 -- # ver2[v]=2 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@368 -- # return 0 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:34:52.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:52.804 --rc genhtml_branch_coverage=1 00:34:52.804 --rc genhtml_function_coverage=1 00:34:52.804 --rc genhtml_legend=1 00:34:52.804 --rc geninfo_all_blocks=1 00:34:52.804 --rc geninfo_unexecuted_blocks=1 00:34:52.804 00:34:52.804 ' 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:34:52.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:52.804 --rc genhtml_branch_coverage=1 00:34:52.804 --rc genhtml_function_coverage=1 00:34:52.804 --rc genhtml_legend=1 00:34:52.804 --rc geninfo_all_blocks=1 00:34:52.804 --rc geninfo_unexecuted_blocks=1 00:34:52.804 00:34:52.804 ' 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:34:52.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:52.804 --rc genhtml_branch_coverage=1 00:34:52.804 --rc genhtml_function_coverage=1 00:34:52.804 --rc genhtml_legend=1 00:34:52.804 --rc geninfo_all_blocks=1 00:34:52.804 --rc geninfo_unexecuted_blocks=1 00:34:52.804 00:34:52.804 ' 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:34:52.804 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:34:52.804 --rc genhtml_branch_coverage=1 00:34:52.804 --rc genhtml_function_coverage=1 00:34:52.804 --rc genhtml_legend=1 00:34:52.804 --rc geninfo_all_blocks=1 00:34:52.804 --rc geninfo_unexecuted_blocks=1 00:34:52.804 00:34:52.804 ' 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@7 -- # uname -s 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@15 -- # shopt -s extglob 00:34:52.804 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@5 -- # export PATH 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@51 -- # : 0 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@55 -- # have_pci_nics=0 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@11 -- # MALLOC_BDEV_SIZE=64 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@13 -- # LVOL_BDEV_INIT_SIZE=20 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@14 -- # LVOL_BDEV_FINAL_SIZE=30 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@18 -- # nvmftestinit 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@474 -- # prepare_net_devs 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@436 -- # local -g is_hw=no 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@438 -- # remove_spdk_ns 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@309 -- # xtrace_disable 00:34:52.805 09:10:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@315 -- # pci_devs=() 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@315 -- # local -a pci_devs 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@316 -- # pci_net_devs=() 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@317 -- # pci_drivers=() 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@317 -- # local -A pci_drivers 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@319 -- # net_devs=() 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@319 -- # local -ga net_devs 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@320 -- # e810=() 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@320 -- # local -ga e810 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@321 -- # x722=() 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@321 -- # local -ga x722 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@322 -- # mlx=() 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@322 -- # local -ga mlx 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:35:00.949 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:35:00.949 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@416 -- # [[ up == up ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:35:00.949 Found net devices under 0000:4b:00.0: cvl_0_0 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@416 -- # [[ up == up ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:35:00.949 Found net devices under 0000:4b:00.1: cvl_0_1 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@440 -- # is_hw=yes 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:35:00.949 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:35:00.950 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:35:00.950 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:35:00.950 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:35:00.950 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:35:00.950 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:35:00.950 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:35:00.950 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:35:00.950 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:35:00.950 09:10:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:35:00.950 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:35:00.950 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.556 ms 00:35:00.950 00:35:00.950 --- 10.0.0.2 ping statistics --- 00:35:00.950 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:00.950 rtt min/avg/max/mdev = 0.556/0.556/0.556/0.000 ms 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:35:00.950 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:35:00.950 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.187 ms 00:35:00.950 00:35:00.950 --- 10.0.0.1 ping statistics --- 00:35:00.950 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:00.950 rtt min/avg/max/mdev = 0.187/0.187/0.187/0.000 ms 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@448 -- # return 0 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@19 -- # nvmfappstart -m 0x7 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@724 -- # xtrace_disable 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@507 -- # nvmfpid=2114539 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@508 -- # waitforlisten 2114539 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x7 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@831 -- # '[' -z 2114539 ']' 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@836 -- # local max_retries=100 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:00.950 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@840 -- # xtrace_disable 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:35:00.950 [2024-11-06 09:10:50.137372] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:35:00.950 [2024-11-06 09:10:50.138502] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:35:00.950 [2024-11-06 09:10:50.138553] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:00.950 [2024-11-06 09:10:50.224512] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:35:00.950 [2024-11-06 09:10:50.266492] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:35:00.950 [2024-11-06 09:10:50.266531] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:35:00.950 [2024-11-06 09:10:50.266539] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:35:00.950 [2024-11-06 09:10:50.266546] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:35:00.950 [2024-11-06 09:10:50.266552] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:35:00.950 [2024-11-06 09:10:50.268025] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:35:00.950 [2024-11-06 09:10:50.268253] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:35:00.950 [2024-11-06 09:10:50.268257] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:00.950 [2024-11-06 09:10:50.326244] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:35:00.950 [2024-11-06 09:10:50.326649] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:35:00.950 [2024-11-06 09:10:50.327039] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:35:00.950 [2024-11-06 09:10:50.327309] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@864 -- # return 0 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@730 -- # xtrace_disable 00:35:00.950 09:10:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:35:00.950 09:10:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:35:00.950 09:10:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:35:01.213 [2024-11-06 09:10:51.172858] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:35:01.213 09:10:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:35:01.474 09:10:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # base_bdevs='Malloc0 ' 00:35:01.474 09:10:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:35:01.735 09:10:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # base_bdevs+=Malloc1 00:35:01.735 09:10:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc0 Malloc1' 00:35:01.736 09:10:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore raid0 lvs 00:35:01.996 09:10:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # lvs=fb85abc4-e2b0-425e-a52f-00d336d6bac7 00:35:01.996 09:10:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u fb85abc4-e2b0-425e-a52f-00d336d6bac7 lvol 20 00:35:02.257 09:10:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # lvol=cfcc0bc0-fe57-475b-9b1f-84d8cd966ebb 00:35:02.257 09:10:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:35:02.257 09:10:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 cfcc0bc0-fe57-475b-9b1f-84d8cd966ebb 00:35:02.518 09:10:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:35:02.779 [2024-11-06 09:10:52.636955] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:02.779 09:10:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:35:02.779 09:10:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@42 -- # perf_pid=2115231 00:35:02.779 09:10:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@44 -- # sleep 1 00:35:02.779 09:10:52 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -o 4096 -q 128 -s 512 -w randwrite -t 10 -c 0x18 00:35:04.161 09:10:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_snapshot cfcc0bc0-fe57-475b-9b1f-84d8cd966ebb MY_SNAPSHOT 00:35:04.161 09:10:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # snapshot=2e5f54a4-6e15-4ae6-9aba-62277bb791da 00:35:04.161 09:10:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_resize cfcc0bc0-fe57-475b-9b1f-84d8cd966ebb 30 00:35:04.422 09:10:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_clone 2e5f54a4-6e15-4ae6-9aba-62277bb791da MY_CLONE 00:35:04.422 09:10:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # clone=310e53cf-d794-42d1-8928-067b040dacb2 00:35:04.422 09:10:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_inflate 310e53cf-d794-42d1-8928-067b040dacb2 00:35:05.074 09:10:54 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@53 -- # wait 2115231 00:35:13.275 Initializing NVMe Controllers 00:35:13.275 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:35:13.275 Controller IO queue size 128, less than required. 00:35:13.275 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:35:13.275 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 3 00:35:13.275 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 4 00:35:13.275 Initialization complete. Launching workers. 00:35:13.275 ======================================================== 00:35:13.275 Latency(us) 00:35:13.275 Device Information : IOPS MiB/s Average min max 00:35:13.275 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 3: 12088.80 47.22 10590.59 1568.50 55033.78 00:35:13.275 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 4: 15032.20 58.72 8517.68 2513.29 65339.29 00:35:13.275 ======================================================== 00:35:13.275 Total : 27121.00 105.94 9441.65 1568.50 65339.29 00:35:13.275 00:35:13.275 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:35:13.275 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete cfcc0bc0-fe57-475b-9b1f-84d8cd966ebb 00:35:13.536 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u fb85abc4-e2b0-425e-a52f-00d336d6bac7 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@60 -- # rm -f 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@62 -- # trap - SIGINT SIGTERM EXIT 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- target/nvmf_lvol.sh@64 -- # nvmftestfini 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@514 -- # nvmfcleanup 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@121 -- # sync 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@124 -- # set +e 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@125 -- # for i in {1..20} 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:35:13.797 rmmod nvme_tcp 00:35:13.797 rmmod nvme_fabrics 00:35:13.797 rmmod nvme_keyring 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@128 -- # set -e 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@129 -- # return 0 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@515 -- # '[' -n 2114539 ']' 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@516 -- # killprocess 2114539 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@950 -- # '[' -z 2114539 ']' 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@954 -- # kill -0 2114539 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@955 -- # uname 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2114539 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2114539' 00:35:13.797 killing process with pid 2114539 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@969 -- # kill 2114539 00:35:13.797 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@974 -- # wait 2114539 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@297 -- # iptr 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@789 -- # iptables-save 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@789 -- # iptables-restore 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@302 -- # remove_spdk_ns 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:14.059 09:11:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:15.975 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:35:15.975 00:35:15.975 real 0m23.548s 00:35:15.975 user 0m55.533s 00:35:15.975 sys 0m10.466s 00:35:15.975 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@1126 -- # xtrace_disable 00:35:15.975 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:35:15.975 ************************************ 00:35:15.975 END TEST nvmf_lvol 00:35:15.975 ************************************ 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@28 -- # run_test nvmf_lvs_grow /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp --interrupt-mode 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:35:16.237 ************************************ 00:35:16.237 START TEST nvmf_lvs_grow 00:35:16.237 ************************************ 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp --interrupt-mode 00:35:16.237 * Looking for test storage... 00:35:16.237 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1689 -- # lcov --version 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@333 -- # local ver1 ver1_l 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@334 -- # local ver2 ver2_l 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@336 -- # IFS=.-: 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@336 -- # read -ra ver1 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@337 -- # IFS=.-: 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@337 -- # read -ra ver2 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@338 -- # local 'op=<' 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@340 -- # ver1_l=2 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@341 -- # ver2_l=1 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@344 -- # case "$op" in 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@345 -- # : 1 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v = 0 )) 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@365 -- # decimal 1 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=1 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 1 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@365 -- # ver1[v]=1 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@366 -- # decimal 2 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@353 -- # local d=2 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@355 -- # echo 2 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@366 -- # ver2[v]=2 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:35:16.237 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@368 -- # return 0 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:35:16.238 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:16.238 --rc genhtml_branch_coverage=1 00:35:16.238 --rc genhtml_function_coverage=1 00:35:16.238 --rc genhtml_legend=1 00:35:16.238 --rc geninfo_all_blocks=1 00:35:16.238 --rc geninfo_unexecuted_blocks=1 00:35:16.238 00:35:16.238 ' 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:35:16.238 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:16.238 --rc genhtml_branch_coverage=1 00:35:16.238 --rc genhtml_function_coverage=1 00:35:16.238 --rc genhtml_legend=1 00:35:16.238 --rc geninfo_all_blocks=1 00:35:16.238 --rc geninfo_unexecuted_blocks=1 00:35:16.238 00:35:16.238 ' 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:35:16.238 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:16.238 --rc genhtml_branch_coverage=1 00:35:16.238 --rc genhtml_function_coverage=1 00:35:16.238 --rc genhtml_legend=1 00:35:16.238 --rc geninfo_all_blocks=1 00:35:16.238 --rc geninfo_unexecuted_blocks=1 00:35:16.238 00:35:16.238 ' 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:35:16.238 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:35:16.238 --rc genhtml_branch_coverage=1 00:35:16.238 --rc genhtml_function_coverage=1 00:35:16.238 --rc genhtml_legend=1 00:35:16.238 --rc geninfo_all_blocks=1 00:35:16.238 --rc geninfo_unexecuted_blocks=1 00:35:16.238 00:35:16.238 ' 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@7 -- # uname -s 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:35:16.238 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@15 -- # shopt -s extglob 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:16.499 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@5 -- # export PATH 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@51 -- # : 0 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@55 -- # have_pci_nics=0 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@12 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@98 -- # nvmftestinit 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@474 -- # prepare_net_devs 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@436 -- # local -g is_hw=no 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@438 -- # remove_spdk_ns 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@309 -- # xtrace_disable 00:35:16.500 09:11:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@315 -- # pci_devs=() 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@315 -- # local -a pci_devs 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@316 -- # pci_net_devs=() 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@317 -- # pci_drivers=() 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@317 -- # local -A pci_drivers 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@319 -- # net_devs=() 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@319 -- # local -ga net_devs 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@320 -- # e810=() 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@320 -- # local -ga e810 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@321 -- # x722=() 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@321 -- # local -ga x722 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@322 -- # mlx=() 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@322 -- # local -ga mlx 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:35:24.640 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:35:24.641 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:35:24.641 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ up == up ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:35:24.641 Found net devices under 0000:4b:00.0: cvl_0_0 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ up == up ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:35:24.641 Found net devices under 0000:4b:00.1: cvl_0_1 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@440 -- # is_hw=yes 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:35:24.641 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:35:24.641 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.662 ms 00:35:24.641 00:35:24.641 --- 10.0.0.2 ping statistics --- 00:35:24.641 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:24.641 rtt min/avg/max/mdev = 0.662/0.662/0.662/0.000 ms 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:35:24.641 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:35:24.641 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.288 ms 00:35:24.641 00:35:24.641 --- 10.0.0.1 ping statistics --- 00:35:24.641 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:35:24.641 rtt min/avg/max/mdev = 0.288/0.288/0.288/0.000 ms 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@448 -- # return 0 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@99 -- # nvmfappstart -m 0x1 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@724 -- # xtrace_disable 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@507 -- # nvmfpid=2121281 00:35:24.641 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@508 -- # waitforlisten 2121281 00:35:24.642 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1 00:35:24.642 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@831 -- # '[' -z 2121281 ']' 00:35:24.642 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:24.642 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@836 -- # local max_retries=100 00:35:24.642 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:24.642 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:24.642 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@840 -- # xtrace_disable 00:35:24.642 09:11:13 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:35:24.642 [2024-11-06 09:11:13.834237] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:35:24.642 [2024-11-06 09:11:13.835528] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:35:24.642 [2024-11-06 09:11:13.835584] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:24.642 [2024-11-06 09:11:13.918064] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:24.642 [2024-11-06 09:11:13.959237] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:35:24.642 [2024-11-06 09:11:13.959273] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:35:24.642 [2024-11-06 09:11:13.959281] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:35:24.642 [2024-11-06 09:11:13.959288] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:35:24.642 [2024-11-06 09:11:13.959294] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:35:24.642 [2024-11-06 09:11:13.959868] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:24.642 [2024-11-06 09:11:14.016718] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:35:24.642 [2024-11-06 09:11:14.016996] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:35:24.642 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:35:24.642 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@864 -- # return 0 00:35:24.642 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:35:24.642 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@730 -- # xtrace_disable 00:35:24.642 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:35:24.642 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:35:24.642 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:35:24.903 [2024-11-06 09:11:14.820617] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@102 -- # run_test lvs_grow_clean lvs_grow 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1107 -- # xtrace_disable 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:35:24.903 ************************************ 00:35:24.903 START TEST lvs_grow_clean 00:35:24.903 ************************************ 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1125 -- # lvs_grow 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:35:24.903 09:11:14 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:35:25.215 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:35:25.215 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:35:25.215 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # lvs=f239901a-72c8-482d-8df0-e9af374a2986 00:35:25.215 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:25.215 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:35:25.475 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:35:25.475 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:35:25.475 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u f239901a-72c8-482d-8df0-e9af374a2986 lvol 150 00:35:25.736 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # lvol=524954a1-502e-4ecb-9f91-0561b726aac2 00:35:25.736 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:35:25.736 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:35:25.736 [2024-11-06 09:11:15.776211] bdev_aio.c:1044:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:35:25.736 [2024-11-06 09:11:15.776291] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:35:25.736 true 00:35:25.736 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:25.736 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:35:25.997 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:35:25.997 09:11:15 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:35:26.256 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 524954a1-502e-4ecb-9f91-0561b726aac2 00:35:26.256 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:35:26.516 [2024-11-06 09:11:16.493038] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:26.516 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:35:26.777 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:35:26.777 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=2121956 00:35:26.777 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:35:26.777 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 2121956 /var/tmp/bdevperf.sock 00:35:26.777 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@831 -- # '[' -z 2121956 ']' 00:35:26.777 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:35:26.777 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@836 -- # local max_retries=100 00:35:26.777 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:35:26.777 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:35:26.777 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@840 -- # xtrace_disable 00:35:26.777 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:35:26.777 [2024-11-06 09:11:16.695391] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:35:26.777 [2024-11-06 09:11:16.695444] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2121956 ] 00:35:26.777 [2024-11-06 09:11:16.781643] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:26.777 [2024-11-06 09:11:16.818374] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:35:27.037 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:35:27.037 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@864 -- # return 0 00:35:27.037 09:11:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:35:27.297 Nvme0n1 00:35:27.297 09:11:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:35:27.558 [ 00:35:27.558 { 00:35:27.558 "name": "Nvme0n1", 00:35:27.558 "aliases": [ 00:35:27.558 "524954a1-502e-4ecb-9f91-0561b726aac2" 00:35:27.558 ], 00:35:27.558 "product_name": "NVMe disk", 00:35:27.558 "block_size": 4096, 00:35:27.558 "num_blocks": 38912, 00:35:27.558 "uuid": "524954a1-502e-4ecb-9f91-0561b726aac2", 00:35:27.558 "numa_id": 0, 00:35:27.558 "assigned_rate_limits": { 00:35:27.558 "rw_ios_per_sec": 0, 00:35:27.558 "rw_mbytes_per_sec": 0, 00:35:27.558 "r_mbytes_per_sec": 0, 00:35:27.558 "w_mbytes_per_sec": 0 00:35:27.558 }, 00:35:27.558 "claimed": false, 00:35:27.558 "zoned": false, 00:35:27.558 "supported_io_types": { 00:35:27.558 "read": true, 00:35:27.558 "write": true, 00:35:27.558 "unmap": true, 00:35:27.558 "flush": true, 00:35:27.558 "reset": true, 00:35:27.558 "nvme_admin": true, 00:35:27.558 "nvme_io": true, 00:35:27.558 "nvme_io_md": false, 00:35:27.558 "write_zeroes": true, 00:35:27.558 "zcopy": false, 00:35:27.558 "get_zone_info": false, 00:35:27.558 "zone_management": false, 00:35:27.558 "zone_append": false, 00:35:27.558 "compare": true, 00:35:27.558 "compare_and_write": true, 00:35:27.558 "abort": true, 00:35:27.558 "seek_hole": false, 00:35:27.558 "seek_data": false, 00:35:27.558 "copy": true, 00:35:27.558 "nvme_iov_md": false 00:35:27.558 }, 00:35:27.558 "memory_domains": [ 00:35:27.558 { 00:35:27.558 "dma_device_id": "system", 00:35:27.558 "dma_device_type": 1 00:35:27.558 } 00:35:27.558 ], 00:35:27.558 "driver_specific": { 00:35:27.558 "nvme": [ 00:35:27.558 { 00:35:27.558 "trid": { 00:35:27.558 "trtype": "TCP", 00:35:27.558 "adrfam": "IPv4", 00:35:27.558 "traddr": "10.0.0.2", 00:35:27.558 "trsvcid": "4420", 00:35:27.558 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:35:27.558 }, 00:35:27.558 "ctrlr_data": { 00:35:27.558 "cntlid": 1, 00:35:27.558 "vendor_id": "0x8086", 00:35:27.558 "model_number": "SPDK bdev Controller", 00:35:27.558 "serial_number": "SPDK0", 00:35:27.558 "firmware_revision": "25.01", 00:35:27.558 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:35:27.558 "oacs": { 00:35:27.558 "security": 0, 00:35:27.558 "format": 0, 00:35:27.558 "firmware": 0, 00:35:27.558 "ns_manage": 0 00:35:27.558 }, 00:35:27.558 "multi_ctrlr": true, 00:35:27.558 "ana_reporting": false 00:35:27.558 }, 00:35:27.558 "vs": { 00:35:27.558 "nvme_version": "1.3" 00:35:27.558 }, 00:35:27.558 "ns_data": { 00:35:27.558 "id": 1, 00:35:27.558 "can_share": true 00:35:27.558 } 00:35:27.558 } 00:35:27.558 ], 00:35:27.558 "mp_policy": "active_passive" 00:35:27.558 } 00:35:27.558 } 00:35:27.558 ] 00:35:27.558 09:11:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=2122016 00:35:27.558 09:11:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:35:27.558 09:11:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:35:27.558 Running I/O for 10 seconds... 00:35:28.499 Latency(us) 00:35:28.499 [2024-11-06T08:11:18.612Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:28.499 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:28.499 Nvme0n1 : 1.00 17680.00 69.06 0.00 0.00 0.00 0.00 0.00 00:35:28.499 [2024-11-06T08:11:18.612Z] =================================================================================================================== 00:35:28.499 [2024-11-06T08:11:18.612Z] Total : 17680.00 69.06 0.00 0.00 0.00 0.00 0.00 00:35:28.499 00:35:29.441 09:11:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:29.441 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:29.441 Nvme0n1 : 2.00 17791.00 69.50 0.00 0.00 0.00 0.00 0.00 00:35:29.441 [2024-11-06T08:11:19.554Z] =================================================================================================================== 00:35:29.441 [2024-11-06T08:11:19.554Z] Total : 17791.00 69.50 0.00 0.00 0.00 0.00 0.00 00:35:29.441 00:35:29.702 true 00:35:29.702 09:11:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:29.702 09:11:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:35:29.703 09:11:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:35:29.703 09:11:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:35:29.703 09:11:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@65 -- # wait 2122016 00:35:30.648 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:30.648 Nvme0n1 : 3.00 17832.33 69.66 0.00 0.00 0.00 0.00 0.00 00:35:30.648 [2024-11-06T08:11:20.761Z] =================================================================================================================== 00:35:30.648 [2024-11-06T08:11:20.761Z] Total : 17832.33 69.66 0.00 0.00 0.00 0.00 0.00 00:35:30.648 00:35:31.589 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:31.589 Nvme0n1 : 4.00 17854.25 69.74 0.00 0.00 0.00 0.00 0.00 00:35:31.589 [2024-11-06T08:11:21.702Z] =================================================================================================================== 00:35:31.589 [2024-11-06T08:11:21.702Z] Total : 17854.25 69.74 0.00 0.00 0.00 0.00 0.00 00:35:31.589 00:35:32.667 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:32.667 Nvme0n1 : 5.00 17880.40 69.85 0.00 0.00 0.00 0.00 0.00 00:35:32.667 [2024-11-06T08:11:22.780Z] =================================================================================================================== 00:35:32.667 [2024-11-06T08:11:22.780Z] Total : 17880.40 69.85 0.00 0.00 0.00 0.00 0.00 00:35:32.667 00:35:33.710 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:33.710 Nvme0n1 : 6.00 17897.50 69.91 0.00 0.00 0.00 0.00 0.00 00:35:33.710 [2024-11-06T08:11:23.823Z] =================================================================================================================== 00:35:33.710 [2024-11-06T08:11:23.823Z] Total : 17897.50 69.91 0.00 0.00 0.00 0.00 0.00 00:35:33.710 00:35:34.652 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:34.652 Nvme0n1 : 7.00 17910.00 69.96 0.00 0.00 0.00 0.00 0.00 00:35:34.652 [2024-11-06T08:11:24.765Z] =================================================================================================================== 00:35:34.652 [2024-11-06T08:11:24.765Z] Total : 17910.00 69.96 0.00 0.00 0.00 0.00 0.00 00:35:34.652 00:35:35.595 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:35.595 Nvme0n1 : 8.00 17919.12 70.00 0.00 0.00 0.00 0.00 0.00 00:35:35.595 [2024-11-06T08:11:25.708Z] =================================================================================================================== 00:35:35.595 [2024-11-06T08:11:25.708Z] Total : 17919.12 70.00 0.00 0.00 0.00 0.00 0.00 00:35:35.595 00:35:36.539 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:36.539 Nvme0n1 : 9.00 17926.00 70.02 0.00 0.00 0.00 0.00 0.00 00:35:36.539 [2024-11-06T08:11:26.652Z] =================================================================================================================== 00:35:36.539 [2024-11-06T08:11:26.652Z] Total : 17926.00 70.02 0.00 0.00 0.00 0.00 0.00 00:35:36.539 00:35:37.481 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:37.481 Nvme0n1 : 10.00 17944.30 70.09 0.00 0.00 0.00 0.00 0.00 00:35:37.481 [2024-11-06T08:11:27.594Z] =================================================================================================================== 00:35:37.481 [2024-11-06T08:11:27.594Z] Total : 17944.30 70.09 0.00 0.00 0.00 0.00 0.00 00:35:37.481 00:35:37.481 00:35:37.481 Latency(us) 00:35:37.481 [2024-11-06T08:11:27.594Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:37.481 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:37.481 Nvme0n1 : 10.01 17946.55 70.10 0.00 0.00 7128.35 2157.23 12997.97 00:35:37.481 [2024-11-06T08:11:27.594Z] =================================================================================================================== 00:35:37.481 [2024-11-06T08:11:27.594Z] Total : 17946.55 70.10 0.00 0.00 7128.35 2157.23 12997.97 00:35:37.481 { 00:35:37.481 "results": [ 00:35:37.481 { 00:35:37.481 "job": "Nvme0n1", 00:35:37.481 "core_mask": "0x2", 00:35:37.481 "workload": "randwrite", 00:35:37.481 "status": "finished", 00:35:37.481 "queue_depth": 128, 00:35:37.481 "io_size": 4096, 00:35:37.481 "runtime": 10.005876, 00:35:37.481 "iops": 17946.554604514386, 00:35:37.481 "mibps": 70.10372892388432, 00:35:37.481 "io_failed": 0, 00:35:37.481 "io_timeout": 0, 00:35:37.481 "avg_latency_us": 7128.3471224195455, 00:35:37.481 "min_latency_us": 2157.2266666666665, 00:35:37.481 "max_latency_us": 12997.973333333333 00:35:37.481 } 00:35:37.481 ], 00:35:37.481 "core_count": 1 00:35:37.481 } 00:35:37.481 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@66 -- # killprocess 2121956 00:35:37.481 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@950 -- # '[' -z 2121956 ']' 00:35:37.481 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@954 -- # kill -0 2121956 00:35:37.481 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@955 -- # uname 00:35:37.481 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:35:37.481 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2121956 00:35:37.742 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:35:37.742 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:35:37.742 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2121956' 00:35:37.742 killing process with pid 2121956 00:35:37.742 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@969 -- # kill 2121956 00:35:37.742 Received shutdown signal, test time was about 10.000000 seconds 00:35:37.742 00:35:37.742 Latency(us) 00:35:37.742 [2024-11-06T08:11:27.855Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:37.742 [2024-11-06T08:11:27.855Z] =================================================================================================================== 00:35:37.742 [2024-11-06T08:11:27.855Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:35:37.742 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@974 -- # wait 2121956 00:35:37.742 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:35:38.003 09:11:27 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:35:38.003 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:38.003 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:35:38.264 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:35:38.264 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@72 -- # [[ '' == \d\i\r\t\y ]] 00:35:38.264 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:35:38.524 [2024-11-06 09:11:28.428227] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@650 -- # local es=0 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:35:38.524 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:38.524 request: 00:35:38.524 { 00:35:38.524 "uuid": "f239901a-72c8-482d-8df0-e9af374a2986", 00:35:38.524 "method": "bdev_lvol_get_lvstores", 00:35:38.524 "req_id": 1 00:35:38.524 } 00:35:38.524 Got JSON-RPC error response 00:35:38.524 response: 00:35:38.524 { 00:35:38.524 "code": -19, 00:35:38.524 "message": "No such device" 00:35:38.524 } 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@653 -- # es=1 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:35:38.785 aio_bdev 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev 524954a1-502e-4ecb-9f91-0561b726aac2 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@899 -- # local bdev_name=524954a1-502e-4ecb-9f91-0561b726aac2 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@901 -- # local i 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:35:38.785 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:35:39.047 09:11:28 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 524954a1-502e-4ecb-9f91-0561b726aac2 -t 2000 00:35:39.047 [ 00:35:39.047 { 00:35:39.047 "name": "524954a1-502e-4ecb-9f91-0561b726aac2", 00:35:39.047 "aliases": [ 00:35:39.047 "lvs/lvol" 00:35:39.047 ], 00:35:39.047 "product_name": "Logical Volume", 00:35:39.047 "block_size": 4096, 00:35:39.047 "num_blocks": 38912, 00:35:39.047 "uuid": "524954a1-502e-4ecb-9f91-0561b726aac2", 00:35:39.047 "assigned_rate_limits": { 00:35:39.047 "rw_ios_per_sec": 0, 00:35:39.047 "rw_mbytes_per_sec": 0, 00:35:39.047 "r_mbytes_per_sec": 0, 00:35:39.047 "w_mbytes_per_sec": 0 00:35:39.047 }, 00:35:39.047 "claimed": false, 00:35:39.047 "zoned": false, 00:35:39.047 "supported_io_types": { 00:35:39.047 "read": true, 00:35:39.047 "write": true, 00:35:39.047 "unmap": true, 00:35:39.047 "flush": false, 00:35:39.047 "reset": true, 00:35:39.047 "nvme_admin": false, 00:35:39.047 "nvme_io": false, 00:35:39.047 "nvme_io_md": false, 00:35:39.047 "write_zeroes": true, 00:35:39.047 "zcopy": false, 00:35:39.047 "get_zone_info": false, 00:35:39.047 "zone_management": false, 00:35:39.047 "zone_append": false, 00:35:39.047 "compare": false, 00:35:39.047 "compare_and_write": false, 00:35:39.047 "abort": false, 00:35:39.047 "seek_hole": true, 00:35:39.047 "seek_data": true, 00:35:39.047 "copy": false, 00:35:39.047 "nvme_iov_md": false 00:35:39.047 }, 00:35:39.047 "driver_specific": { 00:35:39.047 "lvol": { 00:35:39.047 "lvol_store_uuid": "f239901a-72c8-482d-8df0-e9af374a2986", 00:35:39.047 "base_bdev": "aio_bdev", 00:35:39.047 "thin_provision": false, 00:35:39.047 "num_allocated_clusters": 38, 00:35:39.047 "snapshot": false, 00:35:39.047 "clone": false, 00:35:39.047 "esnap_clone": false 00:35:39.047 } 00:35:39.047 } 00:35:39.047 } 00:35:39.047 ] 00:35:39.047 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@907 -- # return 0 00:35:39.047 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:39.047 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:35:39.307 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:35:39.307 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:39.307 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:35:39.568 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:35:39.568 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 524954a1-502e-4ecb-9f91-0561b726aac2 00:35:39.568 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u f239901a-72c8-482d-8df0-e9af374a2986 00:35:39.827 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:35:40.088 09:11:29 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:35:40.088 00:35:40.088 real 0m15.158s 00:35:40.088 user 0m14.746s 00:35:40.088 sys 0m1.330s 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1126 -- # xtrace_disable 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:35:40.088 ************************************ 00:35:40.088 END TEST lvs_grow_clean 00:35:40.088 ************************************ 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@103 -- # run_test lvs_grow_dirty lvs_grow dirty 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1107 -- # xtrace_disable 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:35:40.088 ************************************ 00:35:40.088 START TEST lvs_grow_dirty 00:35:40.088 ************************************ 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1125 -- # lvs_grow dirty 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:35:40.088 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:35:40.349 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:35:40.349 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:35:40.609 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # lvs=5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:40.609 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:35:40.609 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:40.870 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:35:40.870 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:35:40.870 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 lvol 150 00:35:40.870 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # lvol=7ac49db9-f695-4a60-bdd9-44632a46a0de 00:35:40.870 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:35:40.870 09:11:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:35:41.130 [2024-11-06 09:11:31.024190] bdev_aio.c:1044:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:35:41.130 [2024-11-06 09:11:31.024251] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:35:41.130 true 00:35:41.130 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:35:41.131 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:41.131 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:35:41.131 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:35:41.391 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 7ac49db9-f695-4a60-bdd9-44632a46a0de 00:35:41.652 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:35:41.652 [2024-11-06 09:11:31.676454] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:35:41.652 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:35:41.913 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=2124822 00:35:41.913 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:35:41.913 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 2124822 /var/tmp/bdevperf.sock 00:35:41.913 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:35:41.913 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@831 -- # '[' -z 2124822 ']' 00:35:41.913 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:35:41.913 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@836 -- # local max_retries=100 00:35:41.913 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:35:41.913 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:35:41.913 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # xtrace_disable 00:35:41.913 09:11:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:35:41.913 [2024-11-06 09:11:31.908719] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:35:41.913 [2024-11-06 09:11:31.908779] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2124822 ] 00:35:41.913 [2024-11-06 09:11:31.993157] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:41.913 [2024-11-06 09:11:32.022964] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:35:42.856 09:11:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:35:42.856 09:11:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # return 0 00:35:42.856 09:11:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:35:42.856 Nvme0n1 00:35:42.856 09:11:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:35:43.117 [ 00:35:43.117 { 00:35:43.117 "name": "Nvme0n1", 00:35:43.117 "aliases": [ 00:35:43.117 "7ac49db9-f695-4a60-bdd9-44632a46a0de" 00:35:43.117 ], 00:35:43.117 "product_name": "NVMe disk", 00:35:43.117 "block_size": 4096, 00:35:43.117 "num_blocks": 38912, 00:35:43.117 "uuid": "7ac49db9-f695-4a60-bdd9-44632a46a0de", 00:35:43.117 "numa_id": 0, 00:35:43.117 "assigned_rate_limits": { 00:35:43.117 "rw_ios_per_sec": 0, 00:35:43.117 "rw_mbytes_per_sec": 0, 00:35:43.117 "r_mbytes_per_sec": 0, 00:35:43.117 "w_mbytes_per_sec": 0 00:35:43.117 }, 00:35:43.117 "claimed": false, 00:35:43.117 "zoned": false, 00:35:43.117 "supported_io_types": { 00:35:43.117 "read": true, 00:35:43.117 "write": true, 00:35:43.117 "unmap": true, 00:35:43.117 "flush": true, 00:35:43.117 "reset": true, 00:35:43.117 "nvme_admin": true, 00:35:43.117 "nvme_io": true, 00:35:43.117 "nvme_io_md": false, 00:35:43.117 "write_zeroes": true, 00:35:43.117 "zcopy": false, 00:35:43.117 "get_zone_info": false, 00:35:43.117 "zone_management": false, 00:35:43.117 "zone_append": false, 00:35:43.117 "compare": true, 00:35:43.117 "compare_and_write": true, 00:35:43.117 "abort": true, 00:35:43.117 "seek_hole": false, 00:35:43.117 "seek_data": false, 00:35:43.117 "copy": true, 00:35:43.117 "nvme_iov_md": false 00:35:43.117 }, 00:35:43.117 "memory_domains": [ 00:35:43.117 { 00:35:43.117 "dma_device_id": "system", 00:35:43.117 "dma_device_type": 1 00:35:43.117 } 00:35:43.117 ], 00:35:43.117 "driver_specific": { 00:35:43.117 "nvme": [ 00:35:43.117 { 00:35:43.117 "trid": { 00:35:43.117 "trtype": "TCP", 00:35:43.117 "adrfam": "IPv4", 00:35:43.117 "traddr": "10.0.0.2", 00:35:43.117 "trsvcid": "4420", 00:35:43.117 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:35:43.117 }, 00:35:43.117 "ctrlr_data": { 00:35:43.117 "cntlid": 1, 00:35:43.117 "vendor_id": "0x8086", 00:35:43.117 "model_number": "SPDK bdev Controller", 00:35:43.117 "serial_number": "SPDK0", 00:35:43.117 "firmware_revision": "25.01", 00:35:43.117 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:35:43.117 "oacs": { 00:35:43.117 "security": 0, 00:35:43.117 "format": 0, 00:35:43.117 "firmware": 0, 00:35:43.117 "ns_manage": 0 00:35:43.117 }, 00:35:43.117 "multi_ctrlr": true, 00:35:43.117 "ana_reporting": false 00:35:43.117 }, 00:35:43.117 "vs": { 00:35:43.117 "nvme_version": "1.3" 00:35:43.117 }, 00:35:43.117 "ns_data": { 00:35:43.117 "id": 1, 00:35:43.117 "can_share": true 00:35:43.117 } 00:35:43.117 } 00:35:43.117 ], 00:35:43.117 "mp_policy": "active_passive" 00:35:43.117 } 00:35:43.117 } 00:35:43.117 ] 00:35:43.117 09:11:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=2125057 00:35:43.117 09:11:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:35:43.117 09:11:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:35:43.117 Running I/O for 10 seconds... 00:35:44.502 Latency(us) 00:35:44.502 [2024-11-06T08:11:34.615Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:44.502 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:44.502 Nvme0n1 : 1.00 17665.00 69.00 0.00 0.00 0.00 0.00 0.00 00:35:44.502 [2024-11-06T08:11:34.615Z] =================================================================================================================== 00:35:44.502 [2024-11-06T08:11:34.616Z] Total : 17665.00 69.00 0.00 0.00 0.00 0.00 0.00 00:35:44.503 00:35:45.074 09:11:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:45.074 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:45.074 Nvme0n1 : 2.00 17760.00 69.38 0.00 0.00 0.00 0.00 0.00 00:35:45.074 [2024-11-06T08:11:35.187Z] =================================================================================================================== 00:35:45.074 [2024-11-06T08:11:35.187Z] Total : 17760.00 69.38 0.00 0.00 0.00 0.00 0.00 00:35:45.074 00:35:45.336 true 00:35:45.336 09:11:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:45.336 09:11:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:35:45.336 09:11:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:35:45.336 09:11:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:35:45.597 09:11:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@65 -- # wait 2125057 00:35:46.169 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:46.169 Nvme0n1 : 3.00 17792.33 69.50 0.00 0.00 0.00 0.00 0.00 00:35:46.169 [2024-11-06T08:11:36.282Z] =================================================================================================================== 00:35:46.169 [2024-11-06T08:11:36.282Z] Total : 17792.33 69.50 0.00 0.00 0.00 0.00 0.00 00:35:46.169 00:35:47.112 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:47.112 Nvme0n1 : 4.00 17824.25 69.63 0.00 0.00 0.00 0.00 0.00 00:35:47.112 [2024-11-06T08:11:37.225Z] =================================================================================================================== 00:35:47.112 [2024-11-06T08:11:37.225Z] Total : 17824.25 69.63 0.00 0.00 0.00 0.00 0.00 00:35:47.112 00:35:48.497 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:48.497 Nvme0n1 : 5.00 17843.40 69.70 0.00 0.00 0.00 0.00 0.00 00:35:48.497 [2024-11-06T08:11:38.610Z] =================================================================================================================== 00:35:48.497 [2024-11-06T08:11:38.610Z] Total : 17843.40 69.70 0.00 0.00 0.00 0.00 0.00 00:35:48.497 00:35:49.439 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:49.439 Nvme0n1 : 6.00 17866.67 69.79 0.00 0.00 0.00 0.00 0.00 00:35:49.439 [2024-11-06T08:11:39.552Z] =================================================================================================================== 00:35:49.439 [2024-11-06T08:11:39.552Z] Total : 17866.67 69.79 0.00 0.00 0.00 0.00 0.00 00:35:49.439 00:35:50.380 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:50.380 Nvme0n1 : 7.00 17874.29 69.82 0.00 0.00 0.00 0.00 0.00 00:35:50.380 [2024-11-06T08:11:40.493Z] =================================================================================================================== 00:35:50.380 [2024-11-06T08:11:40.493Z] Total : 17874.29 69.82 0.00 0.00 0.00 0.00 0.00 00:35:50.380 00:35:51.322 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:51.322 Nvme0n1 : 8.00 17888.12 69.88 0.00 0.00 0.00 0.00 0.00 00:35:51.322 [2024-11-06T08:11:41.435Z] =================================================================================================================== 00:35:51.322 [2024-11-06T08:11:41.435Z] Total : 17888.12 69.88 0.00 0.00 0.00 0.00 0.00 00:35:51.322 00:35:52.264 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:52.264 Nvme0n1 : 9.00 17898.67 69.92 0.00 0.00 0.00 0.00 0.00 00:35:52.264 [2024-11-06T08:11:42.377Z] =================================================================================================================== 00:35:52.264 [2024-11-06T08:11:42.377Z] Total : 17898.67 69.92 0.00 0.00 0.00 0.00 0.00 00:35:52.264 00:35:53.206 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:53.206 Nvme0n1 : 10.00 17907.30 69.95 0.00 0.00 0.00 0.00 0.00 00:35:53.206 [2024-11-06T08:11:43.319Z] =================================================================================================================== 00:35:53.206 [2024-11-06T08:11:43.319Z] Total : 17907.30 69.95 0.00 0.00 0.00 0.00 0.00 00:35:53.206 00:35:53.206 00:35:53.206 Latency(us) 00:35:53.206 [2024-11-06T08:11:43.319Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:53.206 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:35:53.206 Nvme0n1 : 10.01 17909.91 69.96 0.00 0.00 7143.51 1597.44 12779.52 00:35:53.206 [2024-11-06T08:11:43.319Z] =================================================================================================================== 00:35:53.206 [2024-11-06T08:11:43.319Z] Total : 17909.91 69.96 0.00 0.00 7143.51 1597.44 12779.52 00:35:53.206 { 00:35:53.206 "results": [ 00:35:53.206 { 00:35:53.206 "job": "Nvme0n1", 00:35:53.206 "core_mask": "0x2", 00:35:53.206 "workload": "randwrite", 00:35:53.206 "status": "finished", 00:35:53.206 "queue_depth": 128, 00:35:53.206 "io_size": 4096, 00:35:53.206 "runtime": 10.005691, 00:35:53.206 "iops": 17909.90747165788, 00:35:53.206 "mibps": 69.9605760611636, 00:35:53.206 "io_failed": 0, 00:35:53.206 "io_timeout": 0, 00:35:53.206 "avg_latency_us": 7143.507698580551, 00:35:53.206 "min_latency_us": 1597.44, 00:35:53.206 "max_latency_us": 12779.52 00:35:53.206 } 00:35:53.206 ], 00:35:53.206 "core_count": 1 00:35:53.206 } 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@66 -- # killprocess 2124822 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@950 -- # '[' -z 2124822 ']' 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@954 -- # kill -0 2124822 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@955 -- # uname 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2124822 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2124822' 00:35:53.206 killing process with pid 2124822 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@969 -- # kill 2124822 00:35:53.206 Received shutdown signal, test time was about 10.000000 seconds 00:35:53.206 00:35:53.206 Latency(us) 00:35:53.206 [2024-11-06T08:11:43.319Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:35:53.206 [2024-11-06T08:11:43.319Z] =================================================================================================================== 00:35:53.206 [2024-11-06T08:11:43.319Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:35:53.206 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@974 -- # wait 2124822 00:35:53.467 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:35:53.467 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:35:53.728 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:53.728 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@72 -- # [[ dirty == \d\i\r\t\y ]] 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@74 -- # kill -9 2121281 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # wait 2121281 00:35:53.990 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh: line 75: 2121281 Killed "${NVMF_APP[@]}" "$@" 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # true 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@76 -- # nvmfappstart -m 0x1 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@724 -- # xtrace_disable 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@507 -- # nvmfpid=2127075 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x1 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@508 -- # waitforlisten 2127075 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@831 -- # '[' -z 2127075 ']' 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@836 -- # local max_retries=100 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:35:53.990 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@840 -- # xtrace_disable 00:35:53.990 09:11:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:35:53.990 [2024-11-06 09:11:44.005081] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:35:53.990 [2024-11-06 09:11:44.006105] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:35:53.990 [2024-11-06 09:11:44.006153] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:35:53.990 [2024-11-06 09:11:44.085267] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:35:54.251 [2024-11-06 09:11:44.121906] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:35:54.251 [2024-11-06 09:11:44.121939] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:35:54.251 [2024-11-06 09:11:44.121949] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:35:54.251 [2024-11-06 09:11:44.121957] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:35:54.251 [2024-11-06 09:11:44.121963] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:35:54.251 [2024-11-06 09:11:44.122519] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:35:54.251 [2024-11-06 09:11:44.178395] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:35:54.251 [2024-11-06 09:11:44.178650] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:35:54.825 09:11:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:35:54.825 09:11:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@864 -- # return 0 00:35:54.825 09:11:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:35:54.825 09:11:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@730 -- # xtrace_disable 00:35:54.825 09:11:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:35:54.825 09:11:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:35:54.825 09:11:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:35:55.086 [2024-11-06 09:11:44.997915] blobstore.c:4875:bs_recover: *NOTICE*: Performing recovery on blobstore 00:35:55.086 [2024-11-06 09:11:44.998012] blobstore.c:4822:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x0 00:35:55.086 [2024-11-06 09:11:44.998044] blobstore.c:4822:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x1 00:35:55.086 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # aio_bdev=aio_bdev 00:35:55.086 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@78 -- # waitforbdev 7ac49db9-f695-4a60-bdd9-44632a46a0de 00:35:55.086 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@899 -- # local bdev_name=7ac49db9-f695-4a60-bdd9-44632a46a0de 00:35:55.086 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:35:55.086 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@901 -- # local i 00:35:55.086 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:35:55.086 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:35:55.086 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:35:55.348 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 7ac49db9-f695-4a60-bdd9-44632a46a0de -t 2000 00:35:55.348 [ 00:35:55.348 { 00:35:55.348 "name": "7ac49db9-f695-4a60-bdd9-44632a46a0de", 00:35:55.348 "aliases": [ 00:35:55.348 "lvs/lvol" 00:35:55.348 ], 00:35:55.348 "product_name": "Logical Volume", 00:35:55.348 "block_size": 4096, 00:35:55.348 "num_blocks": 38912, 00:35:55.348 "uuid": "7ac49db9-f695-4a60-bdd9-44632a46a0de", 00:35:55.348 "assigned_rate_limits": { 00:35:55.348 "rw_ios_per_sec": 0, 00:35:55.348 "rw_mbytes_per_sec": 0, 00:35:55.348 "r_mbytes_per_sec": 0, 00:35:55.348 "w_mbytes_per_sec": 0 00:35:55.348 }, 00:35:55.348 "claimed": false, 00:35:55.348 "zoned": false, 00:35:55.348 "supported_io_types": { 00:35:55.348 "read": true, 00:35:55.348 "write": true, 00:35:55.348 "unmap": true, 00:35:55.348 "flush": false, 00:35:55.348 "reset": true, 00:35:55.348 "nvme_admin": false, 00:35:55.348 "nvme_io": false, 00:35:55.348 "nvme_io_md": false, 00:35:55.348 "write_zeroes": true, 00:35:55.348 "zcopy": false, 00:35:55.348 "get_zone_info": false, 00:35:55.348 "zone_management": false, 00:35:55.348 "zone_append": false, 00:35:55.348 "compare": false, 00:35:55.348 "compare_and_write": false, 00:35:55.348 "abort": false, 00:35:55.348 "seek_hole": true, 00:35:55.348 "seek_data": true, 00:35:55.348 "copy": false, 00:35:55.348 "nvme_iov_md": false 00:35:55.348 }, 00:35:55.348 "driver_specific": { 00:35:55.348 "lvol": { 00:35:55.348 "lvol_store_uuid": "5eacc35d-2c5a-4d92-b759-b658b6b4e7f4", 00:35:55.348 "base_bdev": "aio_bdev", 00:35:55.348 "thin_provision": false, 00:35:55.348 "num_allocated_clusters": 38, 00:35:55.348 "snapshot": false, 00:35:55.348 "clone": false, 00:35:55.348 "esnap_clone": false 00:35:55.348 } 00:35:55.348 } 00:35:55.348 } 00:35:55.348 ] 00:35:55.348 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@907 -- # return 0 00:35:55.348 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:55.348 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # jq -r '.[0].free_clusters' 00:35:55.610 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # (( free_clusters == 61 )) 00:35:55.611 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:55.611 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # jq -r '.[0].total_data_clusters' 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # (( data_clusters == 99 )) 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:35:55.872 [2024-11-06 09:11:45.890915] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@650 -- # local es=0 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@652 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@638 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@644 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:35:55.872 09:11:45 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@653 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:56.133 request: 00:35:56.133 { 00:35:56.133 "uuid": "5eacc35d-2c5a-4d92-b759-b658b6b4e7f4", 00:35:56.133 "method": "bdev_lvol_get_lvstores", 00:35:56.133 "req_id": 1 00:35:56.133 } 00:35:56.133 Got JSON-RPC error response 00:35:56.133 response: 00:35:56.133 { 00:35:56.133 "code": -19, 00:35:56.133 "message": "No such device" 00:35:56.133 } 00:35:56.133 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@653 -- # es=1 00:35:56.134 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:35:56.134 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:35:56.134 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:35:56.134 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:35:56.394 aio_bdev 00:35:56.394 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev 7ac49db9-f695-4a60-bdd9-44632a46a0de 00:35:56.394 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@899 -- # local bdev_name=7ac49db9-f695-4a60-bdd9-44632a46a0de 00:35:56.394 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # local bdev_timeout= 00:35:56.394 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@901 -- # local i 00:35:56.394 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # [[ -z '' ]] 00:35:56.395 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # bdev_timeout=2000 00:35:56.395 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:35:56.395 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@906 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 7ac49db9-f695-4a60-bdd9-44632a46a0de -t 2000 00:35:56.655 [ 00:35:56.655 { 00:35:56.655 "name": "7ac49db9-f695-4a60-bdd9-44632a46a0de", 00:35:56.655 "aliases": [ 00:35:56.655 "lvs/lvol" 00:35:56.655 ], 00:35:56.655 "product_name": "Logical Volume", 00:35:56.655 "block_size": 4096, 00:35:56.655 "num_blocks": 38912, 00:35:56.655 "uuid": "7ac49db9-f695-4a60-bdd9-44632a46a0de", 00:35:56.655 "assigned_rate_limits": { 00:35:56.655 "rw_ios_per_sec": 0, 00:35:56.655 "rw_mbytes_per_sec": 0, 00:35:56.655 "r_mbytes_per_sec": 0, 00:35:56.655 "w_mbytes_per_sec": 0 00:35:56.655 }, 00:35:56.655 "claimed": false, 00:35:56.655 "zoned": false, 00:35:56.655 "supported_io_types": { 00:35:56.655 "read": true, 00:35:56.655 "write": true, 00:35:56.655 "unmap": true, 00:35:56.655 "flush": false, 00:35:56.655 "reset": true, 00:35:56.655 "nvme_admin": false, 00:35:56.655 "nvme_io": false, 00:35:56.655 "nvme_io_md": false, 00:35:56.655 "write_zeroes": true, 00:35:56.655 "zcopy": false, 00:35:56.655 "get_zone_info": false, 00:35:56.655 "zone_management": false, 00:35:56.655 "zone_append": false, 00:35:56.655 "compare": false, 00:35:56.655 "compare_and_write": false, 00:35:56.655 "abort": false, 00:35:56.655 "seek_hole": true, 00:35:56.655 "seek_data": true, 00:35:56.655 "copy": false, 00:35:56.655 "nvme_iov_md": false 00:35:56.655 }, 00:35:56.655 "driver_specific": { 00:35:56.655 "lvol": { 00:35:56.655 "lvol_store_uuid": "5eacc35d-2c5a-4d92-b759-b658b6b4e7f4", 00:35:56.655 "base_bdev": "aio_bdev", 00:35:56.655 "thin_provision": false, 00:35:56.655 "num_allocated_clusters": 38, 00:35:56.655 "snapshot": false, 00:35:56.655 "clone": false, 00:35:56.655 "esnap_clone": false 00:35:56.655 } 00:35:56.655 } 00:35:56.655 } 00:35:56.655 ] 00:35:56.655 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@907 -- # return 0 00:35:56.655 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:56.655 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:35:56.916 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:35:56.916 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:56.916 09:11:46 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:35:56.916 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:35:56.916 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 7ac49db9-f695-4a60-bdd9-44632a46a0de 00:35:57.176 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 5eacc35d-2c5a-4d92-b759-b658b6b4e7f4 00:35:57.437 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:35:57.437 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:35:57.713 00:35:57.713 real 0m17.439s 00:35:57.713 user 0m35.279s 00:35:57.713 sys 0m2.981s 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1126 -- # xtrace_disable 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:35:57.713 ************************************ 00:35:57.713 END TEST lvs_grow_dirty 00:35:57.713 ************************************ 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # process_shm --id 0 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@808 -- # type=--id 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@809 -- # id=0 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@810 -- # '[' --id = --pid ']' 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # shm_files=nvmf_trace.0 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@816 -- # [[ -z nvmf_trace.0 ]] 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@820 -- # for n in $shm_files 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@821 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:35:57.713 nvmf_trace.0 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@823 -- # return 0 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # nvmftestfini 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@514 -- # nvmfcleanup 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@121 -- # sync 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@124 -- # set +e 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@125 -- # for i in {1..20} 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:35:57.713 rmmod nvme_tcp 00:35:57.713 rmmod nvme_fabrics 00:35:57.713 rmmod nvme_keyring 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@128 -- # set -e 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@129 -- # return 0 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@515 -- # '[' -n 2127075 ']' 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@516 -- # killprocess 2127075 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@950 -- # '[' -z 2127075 ']' 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@954 -- # kill -0 2127075 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@955 -- # uname 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2127075 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2127075' 00:35:57.713 killing process with pid 2127075 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@969 -- # kill 2127075 00:35:57.713 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@974 -- # wait 2127075 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@297 -- # iptr 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@789 -- # iptables-save 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@789 -- # iptables-restore 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@302 -- # remove_spdk_ns 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:35:58.024 09:11:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:35:59.998 09:11:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:35:59.998 00:35:59.998 real 0m43.856s 00:35:59.998 user 0m52.967s 00:35:59.998 sys 0m10.345s 00:35:59.998 09:11:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@1126 -- # xtrace_disable 00:35:59.998 09:11:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:35:59.998 ************************************ 00:35:59.998 END TEST nvmf_lvs_grow 00:35:59.998 ************************************ 00:35:59.998 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@29 -- # run_test nvmf_bdev_io_wait /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp --interrupt-mode 00:35:59.998 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:35:59.998 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:35:59.998 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:35:59.998 ************************************ 00:35:59.998 START TEST nvmf_bdev_io_wait 00:35:59.998 ************************************ 00:35:59.998 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp --interrupt-mode 00:36:00.260 * Looking for test storage... 00:36:00.260 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1689 -- # lcov --version 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # IFS=.-: 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@336 -- # read -ra ver1 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # IFS=.-: 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@337 -- # read -ra ver2 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@338 -- # local 'op=<' 00:36:00.260 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@340 -- # ver1_l=2 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@341 -- # ver2_l=1 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@344 -- # case "$op" in 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@345 -- # : 1 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # decimal 1 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=1 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 1 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@365 -- # ver1[v]=1 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # decimal 2 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@353 -- # local d=2 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@355 -- # echo 2 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@366 -- # ver2[v]=2 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@368 -- # return 0 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:36:00.261 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:00.261 --rc genhtml_branch_coverage=1 00:36:00.261 --rc genhtml_function_coverage=1 00:36:00.261 --rc genhtml_legend=1 00:36:00.261 --rc geninfo_all_blocks=1 00:36:00.261 --rc geninfo_unexecuted_blocks=1 00:36:00.261 00:36:00.261 ' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:36:00.261 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:00.261 --rc genhtml_branch_coverage=1 00:36:00.261 --rc genhtml_function_coverage=1 00:36:00.261 --rc genhtml_legend=1 00:36:00.261 --rc geninfo_all_blocks=1 00:36:00.261 --rc geninfo_unexecuted_blocks=1 00:36:00.261 00:36:00.261 ' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:36:00.261 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:00.261 --rc genhtml_branch_coverage=1 00:36:00.261 --rc genhtml_function_coverage=1 00:36:00.261 --rc genhtml_legend=1 00:36:00.261 --rc geninfo_all_blocks=1 00:36:00.261 --rc geninfo_unexecuted_blocks=1 00:36:00.261 00:36:00.261 ' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:36:00.261 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:00.261 --rc genhtml_branch_coverage=1 00:36:00.261 --rc genhtml_function_coverage=1 00:36:00.261 --rc genhtml_legend=1 00:36:00.261 --rc geninfo_all_blocks=1 00:36:00.261 --rc geninfo_unexecuted_blocks=1 00:36:00.261 00:36:00.261 ' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # uname -s 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@15 -- # shopt -s extglob 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@5 -- # export PATH 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@51 -- # : 0 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@11 -- # MALLOC_BDEV_SIZE=64 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@14 -- # nvmftestinit 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@474 -- # prepare_net_devs 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@436 -- # local -g is_hw=no 00:36:00.261 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@438 -- # remove_spdk_ns 00:36:00.262 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:00.262 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:00.262 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:00.262 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:36:00.262 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:36:00.262 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@309 -- # xtrace_disable 00:36:00.262 09:11:50 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # pci_devs=() 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # net_devs=() 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # e810=() 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # local -ga e810 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # x722=() 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # local -ga x722 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # mlx=() 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@322 -- # local -ga mlx 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:36:08.409 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:36:08.409 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:08.409 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ up == up ]] 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:36:08.410 Found net devices under 0000:4b:00.0: cvl_0_0 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ up == up ]] 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:36:08.410 Found net devices under 0000:4b:00.1: cvl_0_1 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@440 -- # is_hw=yes 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:08.410 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:08.410 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.592 ms 00:36:08.410 00:36:08.410 --- 10.0.0.2 ping statistics --- 00:36:08.410 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:08.410 rtt min/avg/max/mdev = 0.592/0.592/0.592/0.000 ms 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:08.410 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:08.410 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.311 ms 00:36:08.410 00:36:08.410 --- 10.0.0.1 ping statistics --- 00:36:08.410 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:08.410 rtt min/avg/max/mdev = 0.311/0.311/0.311/0.000 ms 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@448 -- # return 0 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@15 -- # nvmfappstart -m 0xF --wait-for-rpc 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@724 -- # xtrace_disable 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@507 -- # nvmfpid=2132138 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@508 -- # waitforlisten 2132138 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF --wait-for-rpc 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@831 -- # '[' -z 2132138 ']' 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@836 -- # local max_retries=100 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:08.410 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@840 -- # xtrace_disable 00:36:08.410 09:11:57 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.410 [2024-11-06 09:11:57.612158] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:36:08.410 [2024-11-06 09:11:57.613591] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:36:08.410 [2024-11-06 09:11:57.613656] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:36:08.410 [2024-11-06 09:11:57.696663] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:36:08.410 [2024-11-06 09:11:57.739703] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:36:08.410 [2024-11-06 09:11:57.739740] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:36:08.410 [2024-11-06 09:11:57.739754] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:36:08.410 [2024-11-06 09:11:57.739761] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:36:08.410 [2024-11-06 09:11:57.739767] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:36:08.410 [2024-11-06 09:11:57.741521] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:08.410 [2024-11-06 09:11:57.741638] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:36:08.410 [2024-11-06 09:11:57.741796] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:08.410 [2024-11-06 09:11:57.741796] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:36:08.410 [2024-11-06 09:11:57.742071] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:36:08.410 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:36:08.410 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@864 -- # return 0 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@730 -- # xtrace_disable 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@18 -- # rpc_cmd bdev_set_options -p 5 -c 1 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@19 -- # rpc_cmd framework_start_init 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.411 [2024-11-06 09:11:58.484009] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:36:08.411 [2024-11-06 09:11:58.484427] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:36:08.411 [2024-11-06 09:11:58.485153] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:36:08.411 [2024-11-06 09:11:58.485304] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.411 [2024-11-06 09:11:58.494546] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:08.411 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.673 Malloc0 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:08.673 [2024-11-06 09:11:58.554439] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@28 -- # WRITE_PID=2132268 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x10 -i 1 --json /dev/fd/63 -q 128 -o 4096 -w write -t 1 -s 256 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@30 -- # READ_PID=2132271 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # gen_nvmf_target_json 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:36:08.673 { 00:36:08.673 "params": { 00:36:08.673 "name": "Nvme$subsystem", 00:36:08.673 "trtype": "$TEST_TRANSPORT", 00:36:08.673 "traddr": "$NVMF_FIRST_TARGET_IP", 00:36:08.673 "adrfam": "ipv4", 00:36:08.673 "trsvcid": "$NVMF_PORT", 00:36:08.673 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:36:08.673 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:36:08.673 "hdgst": ${hdgst:-false}, 00:36:08.673 "ddgst": ${ddgst:-false} 00:36:08.673 }, 00:36:08.673 "method": "bdev_nvme_attach_controller" 00:36:08.673 } 00:36:08.673 EOF 00:36:08.673 )") 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@32 -- # FLUSH_PID=2132274 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x20 -i 2 --json /dev/fd/63 -q 128 -o 4096 -w read -t 1 -s 256 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # gen_nvmf_target_json 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@34 -- # UNMAP_PID=2132278 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x40 -i 3 --json /dev/fd/63 -q 128 -o 4096 -w flush -t 1 -s 256 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:36:08.673 { 00:36:08.673 "params": { 00:36:08.673 "name": "Nvme$subsystem", 00:36:08.673 "trtype": "$TEST_TRANSPORT", 00:36:08.673 "traddr": "$NVMF_FIRST_TARGET_IP", 00:36:08.673 "adrfam": "ipv4", 00:36:08.673 "trsvcid": "$NVMF_PORT", 00:36:08.673 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:36:08.673 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:36:08.673 "hdgst": ${hdgst:-false}, 00:36:08.673 "ddgst": ${ddgst:-false} 00:36:08.673 }, 00:36:08.673 "method": "bdev_nvme_attach_controller" 00:36:08.673 } 00:36:08.673 EOF 00:36:08.673 )") 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # gen_nvmf_target_json 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@35 -- # sync 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:36:08.673 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:36:08.673 { 00:36:08.673 "params": { 00:36:08.673 "name": "Nvme$subsystem", 00:36:08.673 "trtype": "$TEST_TRANSPORT", 00:36:08.673 "traddr": "$NVMF_FIRST_TARGET_IP", 00:36:08.673 "adrfam": "ipv4", 00:36:08.673 "trsvcid": "$NVMF_PORT", 00:36:08.673 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:36:08.673 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:36:08.673 "hdgst": ${hdgst:-false}, 00:36:08.674 "ddgst": ${ddgst:-false} 00:36:08.674 }, 00:36:08.674 "method": "bdev_nvme_attach_controller" 00:36:08.674 } 00:36:08.674 EOF 00:36:08.674 )") 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x80 -i 4 --json /dev/fd/63 -q 128 -o 4096 -w unmap -t 1 -s 256 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # gen_nvmf_target_json 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # config=() 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # local subsystem config 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:36:08.674 { 00:36:08.674 "params": { 00:36:08.674 "name": "Nvme$subsystem", 00:36:08.674 "trtype": "$TEST_TRANSPORT", 00:36:08.674 "traddr": "$NVMF_FIRST_TARGET_IP", 00:36:08.674 "adrfam": "ipv4", 00:36:08.674 "trsvcid": "$NVMF_PORT", 00:36:08.674 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:36:08.674 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:36:08.674 "hdgst": ${hdgst:-false}, 00:36:08.674 "ddgst": ${ddgst:-false} 00:36:08.674 }, 00:36:08.674 "method": "bdev_nvme_attach_controller" 00:36:08.674 } 00:36:08.674 EOF 00:36:08.674 )") 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@37 -- # wait 2132268 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@580 -- # cat 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:36:08.674 "params": { 00:36:08.674 "name": "Nvme1", 00:36:08.674 "trtype": "tcp", 00:36:08.674 "traddr": "10.0.0.2", 00:36:08.674 "adrfam": "ipv4", 00:36:08.674 "trsvcid": "4420", 00:36:08.674 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:36:08.674 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:36:08.674 "hdgst": false, 00:36:08.674 "ddgst": false 00:36:08.674 }, 00:36:08.674 "method": "bdev_nvme_attach_controller" 00:36:08.674 }' 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@582 -- # jq . 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:36:08.674 "params": { 00:36:08.674 "name": "Nvme1", 00:36:08.674 "trtype": "tcp", 00:36:08.674 "traddr": "10.0.0.2", 00:36:08.674 "adrfam": "ipv4", 00:36:08.674 "trsvcid": "4420", 00:36:08.674 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:36:08.674 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:36:08.674 "hdgst": false, 00:36:08.674 "ddgst": false 00:36:08.674 }, 00:36:08.674 "method": "bdev_nvme_attach_controller" 00:36:08.674 }' 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:36:08.674 "params": { 00:36:08.674 "name": "Nvme1", 00:36:08.674 "trtype": "tcp", 00:36:08.674 "traddr": "10.0.0.2", 00:36:08.674 "adrfam": "ipv4", 00:36:08.674 "trsvcid": "4420", 00:36:08.674 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:36:08.674 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:36:08.674 "hdgst": false, 00:36:08.674 "ddgst": false 00:36:08.674 }, 00:36:08.674 "method": "bdev_nvme_attach_controller" 00:36:08.674 }' 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@583 -- # IFS=, 00:36:08.674 09:11:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:36:08.674 "params": { 00:36:08.674 "name": "Nvme1", 00:36:08.674 "trtype": "tcp", 00:36:08.674 "traddr": "10.0.0.2", 00:36:08.674 "adrfam": "ipv4", 00:36:08.674 "trsvcid": "4420", 00:36:08.674 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:36:08.674 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:36:08.674 "hdgst": false, 00:36:08.674 "ddgst": false 00:36:08.674 }, 00:36:08.674 "method": "bdev_nvme_attach_controller" 00:36:08.674 }' 00:36:08.674 [2024-11-06 09:11:58.608869] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:36:08.674 [2024-11-06 09:11:58.608925] [ DPDK EAL parameters: bdevperf -c 0x10 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:36:08.674 [2024-11-06 09:11:58.610039] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:36:08.674 [2024-11-06 09:11:58.610085] [ DPDK EAL parameters: bdevperf -c 0x40 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk3 --proc-type=auto ] 00:36:08.674 [2024-11-06 09:11:58.612316] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:36:08.674 [2024-11-06 09:11:58.612366] [ DPDK EAL parameters: bdevperf -c 0x20 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk2 --proc-type=auto ] 00:36:08.674 [2024-11-06 09:11:58.613634] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:36:08.674 [2024-11-06 09:11:58.613681] [ DPDK EAL parameters: bdevperf -c 0x80 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk4 --proc-type=auto ] 00:36:08.674 [2024-11-06 09:11:58.767872] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:08.935 [2024-11-06 09:11:58.797734] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:36:08.935 [2024-11-06 09:11:58.808547] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:08.935 [2024-11-06 09:11:58.837296] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:36:08.935 [2024-11-06 09:11:58.854529] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:08.935 [2024-11-06 09:11:58.883448] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 7 00:36:08.935 [2024-11-06 09:11:58.913872] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:08.935 [2024-11-06 09:11:58.942446] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:36:08.935 Running I/O for 1 seconds... 00:36:08.935 Running I/O for 1 seconds... 00:36:08.935 Running I/O for 1 seconds... 00:36:09.196 Running I/O for 1 seconds... 00:36:10.137 12465.00 IOPS, 48.69 MiB/s 00:36:10.137 Latency(us) 00:36:10.137 [2024-11-06T08:12:00.250Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:36:10.137 Job: Nvme1n1 (Core Mask 0x10, workload: write, depth: 128, IO size: 4096) 00:36:10.137 Nvme1n1 : 1.01 12527.02 48.93 0.00 0.00 10182.23 1925.12 12506.45 00:36:10.137 [2024-11-06T08:12:00.250Z] =================================================================================================================== 00:36:10.137 [2024-11-06T08:12:00.250Z] Total : 12527.02 48.93 0.00 0.00 10182.23 1925.12 12506.45 00:36:10.137 12511.00 IOPS, 48.87 MiB/s 00:36:10.137 Latency(us) 00:36:10.137 [2024-11-06T08:12:00.250Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:36:10.137 Job: Nvme1n1 (Core Mask 0x80, workload: unmap, depth: 128, IO size: 4096) 00:36:10.137 Nvme1n1 : 1.01 12575.59 49.12 0.00 0.00 10145.40 2498.56 14527.15 00:36:10.137 [2024-11-06T08:12:00.250Z] =================================================================================================================== 00:36:10.137 [2024-11-06T08:12:00.250Z] Total : 12575.59 49.12 0.00 0.00 10145.40 2498.56 14527.15 00:36:10.137 15743.00 IOPS, 61.50 MiB/s 00:36:10.137 Latency(us) 00:36:10.137 [2024-11-06T08:12:00.250Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:36:10.137 Job: Nvme1n1 (Core Mask 0x20, workload: read, depth: 128, IO size: 4096) 00:36:10.137 Nvme1n1 : 1.01 15797.01 61.71 0.00 0.00 8082.99 3003.73 12943.36 00:36:10.137 [2024-11-06T08:12:00.250Z] =================================================================================================================== 00:36:10.137 [2024-11-06T08:12:00.250Z] Total : 15797.01 61.71 0.00 0.00 8082.99 3003.73 12943.36 00:36:10.137 187120.00 IOPS, 730.94 MiB/s 00:36:10.137 Latency(us) 00:36:10.137 [2024-11-06T08:12:00.250Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:36:10.137 Job: Nvme1n1 (Core Mask 0x40, workload: flush, depth: 128, IO size: 4096) 00:36:10.137 Nvme1n1 : 1.00 186747.97 729.48 0.00 0.00 681.17 315.73 1979.73 00:36:10.137 [2024-11-06T08:12:00.250Z] =================================================================================================================== 00:36:10.137 [2024-11-06T08:12:00.250Z] Total : 186747.97 729.48 0.00 0.00 681.17 315.73 1979.73 00:36:10.137 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@38 -- # wait 2132271 00:36:10.137 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@39 -- # wait 2132274 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@40 -- # wait 2132278 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@42 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@44 -- # trap - SIGINT SIGTERM EXIT 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@46 -- # nvmftestfini 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@514 -- # nvmfcleanup 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@121 -- # sync 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@124 -- # set +e 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@125 -- # for i in {1..20} 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:36:10.138 rmmod nvme_tcp 00:36:10.138 rmmod nvme_fabrics 00:36:10.138 rmmod nvme_keyring 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@128 -- # set -e 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@129 -- # return 0 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@515 -- # '[' -n 2132138 ']' 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@516 -- # killprocess 2132138 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@950 -- # '[' -z 2132138 ']' 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@954 -- # kill -0 2132138 00:36:10.138 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@955 -- # uname 00:36:10.398 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:36:10.398 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2132138 00:36:10.398 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:36:10.398 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:36:10.398 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2132138' 00:36:10.398 killing process with pid 2132138 00:36:10.398 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@969 -- # kill 2132138 00:36:10.398 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@974 -- # wait 2132138 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@297 -- # iptr 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # iptables-save 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@789 -- # iptables-restore 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@302 -- # remove_spdk_ns 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:10.399 09:12:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:36:12.945 00:36:12.945 real 0m12.442s 00:36:12.945 user 0m13.922s 00:36:12.945 sys 0m7.347s 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@1126 -- # xtrace_disable 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:36:12.945 ************************************ 00:36:12.945 END TEST nvmf_bdev_io_wait 00:36:12.945 ************************************ 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@30 -- # run_test nvmf_queue_depth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp --interrupt-mode 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:36:12.945 ************************************ 00:36:12.945 START TEST nvmf_queue_depth 00:36:12.945 ************************************ 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp --interrupt-mode 00:36:12.945 * Looking for test storage... 00:36:12.945 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1689 -- # lcov --version 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@336 -- # IFS=.-: 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@336 -- # read -ra ver1 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@337 -- # IFS=.-: 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@337 -- # read -ra ver2 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@338 -- # local 'op=<' 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@340 -- # ver1_l=2 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@341 -- # ver2_l=1 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@344 -- # case "$op" in 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@345 -- # : 1 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@365 -- # decimal 1 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=1 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 1 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@365 -- # ver1[v]=1 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@366 -- # decimal 2 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@353 -- # local d=2 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@355 -- # echo 2 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@366 -- # ver2[v]=2 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@368 -- # return 0 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:36:12.945 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:12.945 --rc genhtml_branch_coverage=1 00:36:12.945 --rc genhtml_function_coverage=1 00:36:12.945 --rc genhtml_legend=1 00:36:12.945 --rc geninfo_all_blocks=1 00:36:12.945 --rc geninfo_unexecuted_blocks=1 00:36:12.945 00:36:12.945 ' 00:36:12.945 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:36:12.945 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:12.945 --rc genhtml_branch_coverage=1 00:36:12.945 --rc genhtml_function_coverage=1 00:36:12.945 --rc genhtml_legend=1 00:36:12.945 --rc geninfo_all_blocks=1 00:36:12.946 --rc geninfo_unexecuted_blocks=1 00:36:12.946 00:36:12.946 ' 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:36:12.946 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:12.946 --rc genhtml_branch_coverage=1 00:36:12.946 --rc genhtml_function_coverage=1 00:36:12.946 --rc genhtml_legend=1 00:36:12.946 --rc geninfo_all_blocks=1 00:36:12.946 --rc geninfo_unexecuted_blocks=1 00:36:12.946 00:36:12.946 ' 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:36:12.946 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:12.946 --rc genhtml_branch_coverage=1 00:36:12.946 --rc genhtml_function_coverage=1 00:36:12.946 --rc genhtml_legend=1 00:36:12.946 --rc geninfo_all_blocks=1 00:36:12.946 --rc geninfo_unexecuted_blocks=1 00:36:12.946 00:36:12.946 ' 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@7 -- # uname -s 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@15 -- # shopt -s extglob 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@5 -- # export PATH 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@51 -- # : 0 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@14 -- # MALLOC_BDEV_SIZE=64 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@15 -- # MALLOC_BLOCK_SIZE=512 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@17 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@19 -- # nvmftestinit 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@474 -- # prepare_net_devs 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@436 -- # local -g is_hw=no 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@438 -- # remove_spdk_ns 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@309 -- # xtrace_disable 00:36:12.946 09:12:02 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@315 -- # pci_devs=() 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@319 -- # net_devs=() 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@320 -- # e810=() 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@320 -- # local -ga e810 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@321 -- # x722=() 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@321 -- # local -ga x722 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@322 -- # mlx=() 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@322 -- # local -ga mlx 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:36:21.091 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:36:21.091 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ up == up ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:36:21.091 Found net devices under 0000:4b:00.0: cvl_0_0 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ up == up ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:36:21.091 Found net devices under 0000:4b:00.1: cvl_0_1 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@440 -- # is_hw=yes 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:36:21.091 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:21.092 09:12:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:21.092 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:21.092 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.613 ms 00:36:21.092 00:36:21.092 --- 10.0.0.2 ping statistics --- 00:36:21.092 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:21.092 rtt min/avg/max/mdev = 0.613/0.613/0.613/0.000 ms 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:21.092 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:21.092 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.341 ms 00:36:21.092 00:36:21.092 --- 10.0.0.1 ping statistics --- 00:36:21.092 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:21.092 rtt min/avg/max/mdev = 0.341/0.341/0.341/0.000 ms 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@448 -- # return 0 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@21 -- # nvmfappstart -m 0x2 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@724 -- # xtrace_disable 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@507 -- # nvmfpid=2137267 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@508 -- # waitforlisten 2137267 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x2 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@831 -- # '[' -z 2137267 ']' 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@836 -- # local max_retries=100 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:21.092 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@840 -- # xtrace_disable 00:36:21.092 09:12:10 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:21.092 [2024-11-06 09:12:10.399358] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:36:21.092 [2024-11-06 09:12:10.400489] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:36:21.092 [2024-11-06 09:12:10.400544] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:36:21.092 [2024-11-06 09:12:10.505557] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:21.092 [2024-11-06 09:12:10.555201] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:36:21.092 [2024-11-06 09:12:10.555259] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:36:21.092 [2024-11-06 09:12:10.555268] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:36:21.092 [2024-11-06 09:12:10.555275] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:36:21.092 [2024-11-06 09:12:10.555281] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:36:21.092 [2024-11-06 09:12:10.556059] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:21.092 [2024-11-06 09:12:10.639353] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:36:21.092 [2024-11-06 09:12:10.639640] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@864 -- # return 0 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@730 -- # xtrace_disable 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:21.354 [2024-11-06 09:12:11.264938] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@24 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:21.354 Malloc0 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:21.354 [2024-11-06 09:12:11.345151] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@30 -- # bdevperf_pid=2137582 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@32 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 1024 -o 4096 -w verify -t 10 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@33 -- # waitforlisten 2137582 /var/tmp/bdevperf.sock 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@831 -- # '[' -z 2137582 ']' 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@836 -- # local max_retries=100 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:36:21.354 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@840 -- # xtrace_disable 00:36:21.354 09:12:11 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:21.354 [2024-11-06 09:12:11.400981] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:36:21.354 [2024-11-06 09:12:11.401046] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2137582 ] 00:36:21.614 [2024-11-06 09:12:11.479296] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:21.614 [2024-11-06 09:12:11.521114] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:22.185 09:12:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:36:22.185 09:12:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@864 -- # return 0 00:36:22.185 09:12:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@34 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:36:22.185 09:12:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:22.185 09:12:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:22.444 NVMe0n1 00:36:22.444 09:12:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:22.444 09:12:12 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:36:22.444 Running I/O for 10 seconds... 00:36:24.536 9216.00 IOPS, 36.00 MiB/s [2024-11-06T08:12:15.591Z] 9224.50 IOPS, 36.03 MiB/s [2024-11-06T08:12:16.975Z] 9492.00 IOPS, 37.08 MiB/s [2024-11-06T08:12:17.918Z] 10090.50 IOPS, 39.42 MiB/s [2024-11-06T08:12:18.860Z] 10472.80 IOPS, 40.91 MiB/s [2024-11-06T08:12:19.800Z] 10767.50 IOPS, 42.06 MiB/s [2024-11-06T08:12:20.741Z] 10980.29 IOPS, 42.89 MiB/s [2024-11-06T08:12:21.682Z] 11138.75 IOPS, 43.51 MiB/s [2024-11-06T08:12:22.625Z] 11240.00 IOPS, 43.91 MiB/s [2024-11-06T08:12:22.625Z] 11336.90 IOPS, 44.28 MiB/s 00:36:32.512 Latency(us) 00:36:32.512 [2024-11-06T08:12:22.625Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:36:32.512 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 1024, IO size: 4096) 00:36:32.512 Verification LBA range: start 0x0 length 0x4000 00:36:32.512 NVMe0n1 : 10.06 11361.18 44.38 0.00 0.00 89780.14 19114.67 67283.63 00:36:32.512 [2024-11-06T08:12:22.625Z] =================================================================================================================== 00:36:32.512 [2024-11-06T08:12:22.625Z] Total : 11361.18 44.38 0.00 0.00 89780.14 19114.67 67283.63 00:36:32.512 { 00:36:32.512 "results": [ 00:36:32.512 { 00:36:32.512 "job": "NVMe0n1", 00:36:32.512 "core_mask": "0x1", 00:36:32.512 "workload": "verify", 00:36:32.512 "status": "finished", 00:36:32.512 "verify_range": { 00:36:32.512 "start": 0, 00:36:32.512 "length": 16384 00:36:32.512 }, 00:36:32.512 "queue_depth": 1024, 00:36:32.512 "io_size": 4096, 00:36:32.512 "runtime": 10.059782, 00:36:32.512 "iops": 11361.180590195692, 00:36:32.512 "mibps": 44.37961168045192, 00:36:32.512 "io_failed": 0, 00:36:32.512 "io_timeout": 0, 00:36:32.513 "avg_latency_us": 89780.14124926722, 00:36:32.513 "min_latency_us": 19114.666666666668, 00:36:32.513 "max_latency_us": 67283.62666666666 00:36:32.513 } 00:36:32.513 ], 00:36:32.513 "core_count": 1 00:36:32.513 } 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@39 -- # killprocess 2137582 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@950 -- # '[' -z 2137582 ']' 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@954 -- # kill -0 2137582 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@955 -- # uname 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2137582 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2137582' 00:36:32.774 killing process with pid 2137582 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@969 -- # kill 2137582 00:36:32.774 Received shutdown signal, test time was about 10.000000 seconds 00:36:32.774 00:36:32.774 Latency(us) 00:36:32.774 [2024-11-06T08:12:22.887Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:36:32.774 [2024-11-06T08:12:22.887Z] =================================================================================================================== 00:36:32.774 [2024-11-06T08:12:22.887Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@974 -- # wait 2137582 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- target/queue_depth.sh@43 -- # nvmftestfini 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@514 -- # nvmfcleanup 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@121 -- # sync 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@124 -- # set +e 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@125 -- # for i in {1..20} 00:36:32.774 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:36:32.774 rmmod nvme_tcp 00:36:32.774 rmmod nvme_fabrics 00:36:32.774 rmmod nvme_keyring 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@128 -- # set -e 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@129 -- # return 0 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@515 -- # '[' -n 2137267 ']' 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@516 -- # killprocess 2137267 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@950 -- # '[' -z 2137267 ']' 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@954 -- # kill -0 2137267 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@955 -- # uname 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2137267 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2137267' 00:36:33.035 killing process with pid 2137267 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@969 -- # kill 2137267 00:36:33.035 09:12:22 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@974 -- # wait 2137267 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@297 -- # iptr 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@789 -- # iptables-save 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@789 -- # iptables-restore 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@302 -- # remove_spdk_ns 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:33.035 09:12:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:36:35.580 00:36:35.580 real 0m22.571s 00:36:35.580 user 0m24.907s 00:36:35.580 sys 0m7.397s 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@1126 -- # xtrace_disable 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:36:35.580 ************************************ 00:36:35.580 END TEST nvmf_queue_depth 00:36:35.580 ************************************ 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@31 -- # run_test nvmf_target_multipath /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp --interrupt-mode 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:36:35.580 ************************************ 00:36:35.580 START TEST nvmf_target_multipath 00:36:35.580 ************************************ 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp --interrupt-mode 00:36:35.580 * Looking for test storage... 00:36:35.580 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1689 -- # lcov --version 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@336 -- # IFS=.-: 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@336 -- # read -ra ver1 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@337 -- # IFS=.-: 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@337 -- # read -ra ver2 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@338 -- # local 'op=<' 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@340 -- # ver1_l=2 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@341 -- # ver2_l=1 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@344 -- # case "$op" in 00:36:35.580 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@345 -- # : 1 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@365 -- # decimal 1 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=1 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 1 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@365 -- # ver1[v]=1 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@366 -- # decimal 2 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@353 -- # local d=2 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@355 -- # echo 2 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@366 -- # ver2[v]=2 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@368 -- # return 0 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:36:35.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:35.581 --rc genhtml_branch_coverage=1 00:36:35.581 --rc genhtml_function_coverage=1 00:36:35.581 --rc genhtml_legend=1 00:36:35.581 --rc geninfo_all_blocks=1 00:36:35.581 --rc geninfo_unexecuted_blocks=1 00:36:35.581 00:36:35.581 ' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:36:35.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:35.581 --rc genhtml_branch_coverage=1 00:36:35.581 --rc genhtml_function_coverage=1 00:36:35.581 --rc genhtml_legend=1 00:36:35.581 --rc geninfo_all_blocks=1 00:36:35.581 --rc geninfo_unexecuted_blocks=1 00:36:35.581 00:36:35.581 ' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:36:35.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:35.581 --rc genhtml_branch_coverage=1 00:36:35.581 --rc genhtml_function_coverage=1 00:36:35.581 --rc genhtml_legend=1 00:36:35.581 --rc geninfo_all_blocks=1 00:36:35.581 --rc geninfo_unexecuted_blocks=1 00:36:35.581 00:36:35.581 ' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:36:35.581 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:35.581 --rc genhtml_branch_coverage=1 00:36:35.581 --rc genhtml_function_coverage=1 00:36:35.581 --rc genhtml_legend=1 00:36:35.581 --rc geninfo_all_blocks=1 00:36:35.581 --rc geninfo_unexecuted_blocks=1 00:36:35.581 00:36:35.581 ' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@7 -- # uname -s 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@15 -- # shopt -s extglob 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@5 -- # export PATH 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@51 -- # : 0 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@11 -- # MALLOC_BDEV_SIZE=64 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@43 -- # nvmftestinit 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@474 -- # prepare_net_devs 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@436 -- # local -g is_hw=no 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@438 -- # remove_spdk_ns 00:36:35.581 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:35.582 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:35.582 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:35.582 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:36:35.582 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:36:35.582 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@309 -- # xtrace_disable 00:36:35.582 09:12:25 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:36:43.723 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:43.723 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@315 -- # pci_devs=() 00:36:43.723 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:43.723 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:43.723 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:43.723 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:43.723 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:43.723 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@319 -- # net_devs=() 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@320 -- # e810=() 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@320 -- # local -ga e810 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@321 -- # x722=() 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@321 -- # local -ga x722 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@322 -- # mlx=() 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@322 -- # local -ga mlx 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:36:43.724 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:36:43.724 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ up == up ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:36:43.724 Found net devices under 0000:4b:00.0: cvl_0_0 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ up == up ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:36:43.724 Found net devices under 0000:4b:00.1: cvl_0_1 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@440 -- # is_hw=yes 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:43.724 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:43.725 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:43.725 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.606 ms 00:36:43.725 00:36:43.725 --- 10.0.0.2 ping statistics --- 00:36:43.725 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:43.725 rtt min/avg/max/mdev = 0.606/0.606/0.606/0.000 ms 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:43.725 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:43.725 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.248 ms 00:36:43.725 00:36:43.725 --- 10.0.0.1 ping statistics --- 00:36:43.725 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:43.725 rtt min/avg/max/mdev = 0.248/0.248/0.248/0.000 ms 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@448 -- # return 0 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@45 -- # '[' -z ']' 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@46 -- # echo 'only one NIC for nvmf test' 00:36:43.725 only one NIC for nvmf test 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@47 -- # nvmftestfini 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@514 -- # nvmfcleanup 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:36:43.725 rmmod nvme_tcp 00:36:43.725 rmmod nvme_fabrics 00:36:43.725 rmmod nvme_keyring 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-save 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-restore 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:43.725 09:12:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@48 -- # exit 0 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- target/multipath.sh@1 -- # nvmftestfini 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@514 -- # nvmfcleanup 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@121 -- # sync 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@124 -- # set +e 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@125 -- # for i in {1..20} 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@128 -- # set -e 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@129 -- # return 0 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@515 -- # '[' -n '' ']' 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@297 -- # iptr 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-save 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@789 -- # iptables-restore 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@302 -- # remove_spdk_ns 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:36:45.113 00:36:45.113 real 0m9.585s 00:36:45.113 user 0m2.061s 00:36:45.113 sys 0m5.468s 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@1126 -- # xtrace_disable 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:36:45.113 ************************************ 00:36:45.113 END TEST nvmf_target_multipath 00:36:45.113 ************************************ 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@32 -- # run_test nvmf_zcopy /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp --interrupt-mode 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:36:45.113 ************************************ 00:36:45.113 START TEST nvmf_zcopy 00:36:45.113 ************************************ 00:36:45.113 09:12:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp --interrupt-mode 00:36:45.113 * Looking for test storage... 00:36:45.113 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1689 -- # lcov --version 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@333 -- # local ver1 ver1_l 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@334 -- # local ver2 ver2_l 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@336 -- # IFS=.-: 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@336 -- # read -ra ver1 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@337 -- # IFS=.-: 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@337 -- # read -ra ver2 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@338 -- # local 'op=<' 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@340 -- # ver1_l=2 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@341 -- # ver2_l=1 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@344 -- # case "$op" in 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@345 -- # : 1 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@364 -- # (( v = 0 )) 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@365 -- # decimal 1 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@353 -- # local d=1 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@355 -- # echo 1 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@365 -- # ver1[v]=1 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@366 -- # decimal 2 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@353 -- # local d=2 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@355 -- # echo 2 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@366 -- # ver2[v]=2 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@368 -- # return 0 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:36:45.113 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:45.113 --rc genhtml_branch_coverage=1 00:36:45.113 --rc genhtml_function_coverage=1 00:36:45.113 --rc genhtml_legend=1 00:36:45.113 --rc geninfo_all_blocks=1 00:36:45.113 --rc geninfo_unexecuted_blocks=1 00:36:45.113 00:36:45.113 ' 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:36:45.113 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:45.113 --rc genhtml_branch_coverage=1 00:36:45.113 --rc genhtml_function_coverage=1 00:36:45.113 --rc genhtml_legend=1 00:36:45.113 --rc geninfo_all_blocks=1 00:36:45.113 --rc geninfo_unexecuted_blocks=1 00:36:45.113 00:36:45.113 ' 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:36:45.113 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:45.113 --rc genhtml_branch_coverage=1 00:36:45.113 --rc genhtml_function_coverage=1 00:36:45.113 --rc genhtml_legend=1 00:36:45.113 --rc geninfo_all_blocks=1 00:36:45.113 --rc geninfo_unexecuted_blocks=1 00:36:45.113 00:36:45.113 ' 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:36:45.113 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:36:45.113 --rc genhtml_branch_coverage=1 00:36:45.113 --rc genhtml_function_coverage=1 00:36:45.113 --rc genhtml_legend=1 00:36:45.113 --rc geninfo_all_blocks=1 00:36:45.113 --rc geninfo_unexecuted_blocks=1 00:36:45.113 00:36:45.113 ' 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@7 -- # uname -s 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:36:45.113 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@15 -- # shopt -s extglob 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@5 -- # export PATH 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@51 -- # : 0 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@55 -- # have_pci_nics=0 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@12 -- # nvmftestinit 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@474 -- # prepare_net_devs 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@436 -- # local -g is_hw=no 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@438 -- # remove_spdk_ns 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@309 -- # xtrace_disable 00:36:45.114 09:12:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@315 -- # pci_devs=() 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@315 -- # local -a pci_devs 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@316 -- # pci_net_devs=() 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@317 -- # pci_drivers=() 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@317 -- # local -A pci_drivers 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@319 -- # net_devs=() 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@319 -- # local -ga net_devs 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@320 -- # e810=() 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@320 -- # local -ga e810 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@321 -- # x722=() 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@321 -- # local -ga x722 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@322 -- # mlx=() 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@322 -- # local -ga mlx 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:36:53.263 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:36:53.263 09:12:41 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:36:53.263 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ up == up ]] 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:53.263 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:36:53.263 Found net devices under 0000:4b:00.0: cvl_0_0 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ up == up ]] 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:36:53.264 Found net devices under 0000:4b:00.1: cvl_0_1 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@440 -- # is_hw=yes 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:36:53.264 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:36:53.264 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.605 ms 00:36:53.264 00:36:53.264 --- 10.0.0.2 ping statistics --- 00:36:53.264 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:53.264 rtt min/avg/max/mdev = 0.605/0.605/0.605/0.000 ms 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:36:53.264 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:36:53.264 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.352 ms 00:36:53.264 00:36:53.264 --- 10.0.0.1 ping statistics --- 00:36:53.264 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:36:53.264 rtt min/avg/max/mdev = 0.352/0.352/0.352/0.000 ms 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@448 -- # return 0 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@13 -- # nvmfappstart -m 0x2 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@724 -- # xtrace_disable 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@507 -- # nvmfpid=2148027 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@508 -- # waitforlisten 2148027 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x2 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@831 -- # '[' -z 2148027 ']' 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@836 -- # local max_retries=100 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:36:53.264 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@840 -- # xtrace_disable 00:36:53.264 09:12:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:36:53.264 [2024-11-06 09:12:42.400638] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:36:53.264 [2024-11-06 09:12:42.401806] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:36:53.264 [2024-11-06 09:12:42.401859] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:36:53.264 [2024-11-06 09:12:42.501785] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:53.264 [2024-11-06 09:12:42.551318] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:36:53.264 [2024-11-06 09:12:42.551369] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:36:53.264 [2024-11-06 09:12:42.551377] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:36:53.264 [2024-11-06 09:12:42.551384] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:36:53.264 [2024-11-06 09:12:42.551391] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:36:53.264 [2024-11-06 09:12:42.552162] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:36:53.264 [2024-11-06 09:12:42.629477] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:36:53.264 [2024-11-06 09:12:42.629771] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@864 -- # return 0 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@730 -- # xtrace_disable 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@15 -- # '[' tcp '!=' tcp ']' 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@22 -- # rpc_cmd nvmf_create_transport -t tcp -o -c 0 --zcopy 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:36:53.264 [2024-11-06 09:12:43.248991] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:36:53.264 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:36:53.265 [2024-11-06 09:12:43.277270] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@29 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc0 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:36:53.265 malloc0 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -t 10 -q 128 -w verify -o 8192 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@33 -- # gen_nvmf_target_json 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@558 -- # config=() 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@558 -- # local subsystem config 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:36:53.265 { 00:36:53.265 "params": { 00:36:53.265 "name": "Nvme$subsystem", 00:36:53.265 "trtype": "$TEST_TRANSPORT", 00:36:53.265 "traddr": "$NVMF_FIRST_TARGET_IP", 00:36:53.265 "adrfam": "ipv4", 00:36:53.265 "trsvcid": "$NVMF_PORT", 00:36:53.265 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:36:53.265 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:36:53.265 "hdgst": ${hdgst:-false}, 00:36:53.265 "ddgst": ${ddgst:-false} 00:36:53.265 }, 00:36:53.265 "method": "bdev_nvme_attach_controller" 00:36:53.265 } 00:36:53.265 EOF 00:36:53.265 )") 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@580 -- # cat 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # jq . 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@583 -- # IFS=, 00:36:53.265 09:12:43 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:36:53.265 "params": { 00:36:53.265 "name": "Nvme1", 00:36:53.265 "trtype": "tcp", 00:36:53.265 "traddr": "10.0.0.2", 00:36:53.265 "adrfam": "ipv4", 00:36:53.265 "trsvcid": "4420", 00:36:53.265 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:36:53.265 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:36:53.265 "hdgst": false, 00:36:53.265 "ddgst": false 00:36:53.265 }, 00:36:53.265 "method": "bdev_nvme_attach_controller" 00:36:53.265 }' 00:36:53.525 [2024-11-06 09:12:43.380640] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:36:53.525 [2024-11-06 09:12:43.380712] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2148124 ] 00:36:53.525 [2024-11-06 09:12:43.458236] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:36:53.525 [2024-11-06 09:12:43.499914] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:36:53.786 Running I/O for 10 seconds... 00:36:55.668 6519.00 IOPS, 50.93 MiB/s [2024-11-06T08:12:47.165Z] 6566.00 IOPS, 51.30 MiB/s [2024-11-06T08:12:48.106Z] 6587.00 IOPS, 51.46 MiB/s [2024-11-06T08:12:49.047Z] 6598.25 IOPS, 51.55 MiB/s [2024-11-06T08:12:49.989Z] 6603.80 IOPS, 51.59 MiB/s [2024-11-06T08:12:50.932Z] 6998.50 IOPS, 54.68 MiB/s [2024-11-06T08:12:51.919Z] 7361.57 IOPS, 57.51 MiB/s [2024-11-06T08:12:52.888Z] 7637.12 IOPS, 59.67 MiB/s [2024-11-06T08:12:53.831Z] 7851.78 IOPS, 61.34 MiB/s [2024-11-06T08:12:53.831Z] 8022.70 IOPS, 62.68 MiB/s 00:37:03.718 Latency(us) 00:37:03.718 [2024-11-06T08:12:53.831Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:03.718 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 8192) 00:37:03.718 Verification LBA range: start 0x0 length 0x1000 00:37:03.718 Nvme1n1 : 10.01 8026.04 62.70 0.00 0.00 15894.31 1829.55 27525.12 00:37:03.718 [2024-11-06T08:12:53.831Z] =================================================================================================================== 00:37:03.718 [2024-11-06T08:12:53.831Z] Total : 8026.04 62.70 0.00 0.00 15894.31 1829.55 27525.12 00:37:03.980 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@39 -- # perfpid=2150130 00:37:03.980 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@41 -- # xtrace_disable 00:37:03.980 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:37:03.980 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -t 5 -q 128 -w randrw -M 50 -o 8192 00:37:03.980 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@37 -- # gen_nvmf_target_json 00:37:03.980 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@558 -- # config=() 00:37:03.980 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@558 -- # local subsystem config 00:37:03.980 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:37:03.980 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:37:03.980 { 00:37:03.980 "params": { 00:37:03.980 "name": "Nvme$subsystem", 00:37:03.980 "trtype": "$TEST_TRANSPORT", 00:37:03.981 "traddr": "$NVMF_FIRST_TARGET_IP", 00:37:03.981 "adrfam": "ipv4", 00:37:03.981 "trsvcid": "$NVMF_PORT", 00:37:03.981 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:37:03.981 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:37:03.981 "hdgst": ${hdgst:-false}, 00:37:03.981 "ddgst": ${ddgst:-false} 00:37:03.981 }, 00:37:03.981 "method": "bdev_nvme_attach_controller" 00:37:03.981 } 00:37:03.981 EOF 00:37:03.981 )") 00:37:03.981 [2024-11-06 09:12:53.920558] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:53.920588] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@580 -- # cat 00:37:03.981 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@582 -- # jq . 00:37:03.981 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@583 -- # IFS=, 00:37:03.981 09:12:53 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:37:03.981 "params": { 00:37:03.981 "name": "Nvme1", 00:37:03.981 "trtype": "tcp", 00:37:03.981 "traddr": "10.0.0.2", 00:37:03.981 "adrfam": "ipv4", 00:37:03.981 "trsvcid": "4420", 00:37:03.981 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:37:03.981 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:37:03.981 "hdgst": false, 00:37:03.981 "ddgst": false 00:37:03.981 }, 00:37:03.981 "method": "bdev_nvme_attach_controller" 00:37:03.981 }' 00:37:03.981 [2024-11-06 09:12:53.932529] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:53.932539] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:53.944527] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:53.944536] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:53.956527] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:53.956535] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:53.966929] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:37:03.981 [2024-11-06 09:12:53.966978] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2150130 ] 00:37:03.981 [2024-11-06 09:12:53.968526] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:53.968535] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:53.980527] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:53.980535] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:53.992527] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:53.992535] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:54.004526] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:54.004534] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:54.016526] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:54.016535] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:54.028526] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:54.028534] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:54.036640] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:37:03.981 [2024-11-06 09:12:54.040527] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:54.040535] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:54.052541] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:54.052551] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:54.064528] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:54.064537] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:54.072577] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:37:03.981 [2024-11-06 09:12:54.076528] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:54.076538] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:03.981 [2024-11-06 09:12:54.088532] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:03.981 [2024-11-06 09:12:54.088544] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.100530] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.100545] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.112528] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.112538] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.124528] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.124539] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.136528] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.136536] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.148535] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.148551] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.160530] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.160542] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.172529] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.172539] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.184528] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.184539] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.196528] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.196540] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.208534] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.208551] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 Running I/O for 5 seconds... 00:37:04.242 [2024-11-06 09:12:54.223829] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.223847] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.237433] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.237449] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.251344] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.251361] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.264172] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.264190] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.276492] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.276509] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.289830] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.289846] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.304038] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.304055] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.317140] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.317155] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.331709] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.331730] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.242 [2024-11-06 09:12:54.344627] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.242 [2024-11-06 09:12:54.344643] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.357213] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.357229] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.371668] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.371684] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.384601] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.384616] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.397112] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.397128] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.412002] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.412018] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.424740] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.424759] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.439601] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.439617] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.452689] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.452706] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.464255] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.464271] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.477098] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.477114] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.491839] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.491855] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.504580] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.504 [2024-11-06 09:12:54.504595] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.504 [2024-11-06 09:12:54.516957] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.505 [2024-11-06 09:12:54.516972] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.505 [2024-11-06 09:12:54.531813] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.505 [2024-11-06 09:12:54.531829] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.505 [2024-11-06 09:12:54.545172] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.505 [2024-11-06 09:12:54.545187] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.505 [2024-11-06 09:12:54.559728] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.505 [2024-11-06 09:12:54.559744] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.505 [2024-11-06 09:12:54.573000] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.505 [2024-11-06 09:12:54.573016] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.505 [2024-11-06 09:12:54.587894] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.505 [2024-11-06 09:12:54.587915] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.505 [2024-11-06 09:12:54.601083] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.505 [2024-11-06 09:12:54.601099] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.615847] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.615864] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.628513] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.628529] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.640845] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.640860] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.655802] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.655818] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.668559] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.668575] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.680990] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.681006] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.695527] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.695542] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.708690] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.708706] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.720273] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.720289] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.733394] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.733410] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.747979] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.747995] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.760854] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.760868] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.775332] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.775348] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.788274] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.788290] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.800761] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.800775] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.815926] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.815942] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.829132] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.829147] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.844103] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.844122] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.856408] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.856424] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:04.766 [2024-11-06 09:12:54.869565] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:04.766 [2024-11-06 09:12:54.869581] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:54.884387] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:54.884403] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:54.897179] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:54.897195] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:54.911548] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:54.911563] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:54.924208] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:54.924223] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:54.936737] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:54.936756] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:54.951240] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:54.951256] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:54.964321] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:54.964336] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:54.975980] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:54.975995] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:54.988995] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:54.989010] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:55.003925] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:55.003940] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:55.016856] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:55.016870] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.027 [2024-11-06 09:12:55.031759] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.027 [2024-11-06 09:12:55.031774] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.028 [2024-11-06 09:12:55.044999] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.028 [2024-11-06 09:12:55.045014] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.028 [2024-11-06 09:12:55.059444] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.028 [2024-11-06 09:12:55.059459] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.028 [2024-11-06 09:12:55.071943] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.028 [2024-11-06 09:12:55.071958] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.028 [2024-11-06 09:12:55.084662] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.028 [2024-11-06 09:12:55.084678] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.028 [2024-11-06 09:12:55.096582] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.028 [2024-11-06 09:12:55.096605] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.028 [2024-11-06 09:12:55.109563] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.028 [2024-11-06 09:12:55.109579] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.028 [2024-11-06 09:12:55.123736] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.028 [2024-11-06 09:12:55.123756] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.028 [2024-11-06 09:12:55.136771] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.028 [2024-11-06 09:12:55.136787] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.148843] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.148858] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.163577] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.163592] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.176916] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.176930] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.192033] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.192048] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.204702] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.204717] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.217185] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.217200] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 18857.00 IOPS, 147.32 MiB/s [2024-11-06T08:12:55.401Z] [2024-11-06 09:12:55.231708] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.231724] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.244947] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.244962] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.259783] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.259798] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.272949] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.272964] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.288120] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.288135] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.300660] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.300675] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.315802] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.315817] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.328352] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.288 [2024-11-06 09:12:55.328368] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.288 [2024-11-06 09:12:55.340542] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.289 [2024-11-06 09:12:55.340558] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.289 [2024-11-06 09:12:55.353216] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.289 [2024-11-06 09:12:55.353231] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.289 [2024-11-06 09:12:55.368044] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.289 [2024-11-06 09:12:55.368059] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.289 [2024-11-06 09:12:55.380775] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.289 [2024-11-06 09:12:55.380789] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.289 [2024-11-06 09:12:55.395635] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.289 [2024-11-06 09:12:55.395650] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.408734] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.408754] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.423619] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.423634] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.436856] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.436871] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.451698] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.451713] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.464339] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.464354] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.475965] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.475981] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.488982] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.488997] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.504034] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.504049] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.517287] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.517302] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.531922] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.531939] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.544568] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.544584] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.557419] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.557434] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.571967] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.571982] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.585014] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.585029] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.599959] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.599974] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.613100] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.613115] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.628180] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.628196] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.641020] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.641036] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.550 [2024-11-06 09:12:55.655900] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.550 [2024-11-06 09:12:55.655915] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.668934] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.668949] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.683736] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.683755] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.696763] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.696778] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.708911] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.708925] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.723341] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.723356] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.736547] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.736562] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.749050] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.749066] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.764187] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.764202] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.777344] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.777359] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.791790] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.791806] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.804805] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.804820] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.819800] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.819815] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.832550] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.832565] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.844483] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.844498] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.856424] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.856444] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.810 [2024-11-06 09:12:55.869428] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.810 [2024-11-06 09:12:55.869443] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.811 [2024-11-06 09:12:55.883871] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.811 [2024-11-06 09:12:55.883887] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.811 [2024-11-06 09:12:55.896549] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.811 [2024-11-06 09:12:55.896565] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:05.811 [2024-11-06 09:12:55.908946] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:05.811 [2024-11-06 09:12:55.908960] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:55.923798] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:55.923814] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:55.936832] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:55.936848] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:55.951566] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:55.951581] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:55.964948] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:55.964963] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:55.979752] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:55.979768] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:55.992883] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:55.992899] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:56.008172] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:56.008188] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:56.021136] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:56.021152] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:56.035812] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:56.035828] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:56.049000] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:56.049016] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.071 [2024-11-06 09:12:56.063933] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.071 [2024-11-06 09:12:56.063948] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.072 [2024-11-06 09:12:56.076500] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.072 [2024-11-06 09:12:56.076517] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.072 [2024-11-06 09:12:56.088615] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.072 [2024-11-06 09:12:56.088630] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.072 [2024-11-06 09:12:56.100507] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.072 [2024-11-06 09:12:56.100523] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.072 [2024-11-06 09:12:56.113133] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.072 [2024-11-06 09:12:56.113152] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.072 [2024-11-06 09:12:56.127813] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.072 [2024-11-06 09:12:56.127830] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.072 [2024-11-06 09:12:56.140733] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.072 [2024-11-06 09:12:56.140754] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.072 [2024-11-06 09:12:56.152825] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.072 [2024-11-06 09:12:56.152842] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.072 [2024-11-06 09:12:56.167242] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.072 [2024-11-06 09:12:56.167258] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.072 [2024-11-06 09:12:56.180289] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.072 [2024-11-06 09:12:56.180305] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.193049] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.193064] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.207757] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.207773] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.220488] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.220504] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 18883.00 IOPS, 147.52 MiB/s [2024-11-06T08:12:56.446Z] [2024-11-06 09:12:56.233146] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.233161] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.247607] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.247623] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.260675] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.260691] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.273479] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.273495] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.288360] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.288375] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.300930] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.300945] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.315449] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.315465] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.328027] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.328042] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.340557] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.340573] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.352875] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.352890] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.367785] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.367805] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.380388] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.380404] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.393088] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.393103] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.407928] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.407943] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.420977] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.420992] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.333 [2024-11-06 09:12:56.435900] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.333 [2024-11-06 09:12:56.435916] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.449076] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.449091] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.463888] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.463904] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.476897] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.476912] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.492020] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.492036] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.505105] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.505121] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.519757] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.519772] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.532683] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.532698] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.544722] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.544737] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.557302] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.557317] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.571581] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.571596] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.584758] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.584773] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.599839] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.599855] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.612453] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.612469] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.624606] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.624621] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.637303] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.637318] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.651644] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.651660] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.664601] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.664615] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.676253] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.676268] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.689126] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.689141] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.594 [2024-11-06 09:12:56.703721] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.594 [2024-11-06 09:12:56.703736] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.716399] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.716415] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.729162] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.729177] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.743714] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.743731] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.756420] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.756436] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.769010] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.769025] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.784191] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.784206] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.796985] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.797000] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.811684] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.811699] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.824831] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.824845] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.839283] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.839298] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.852485] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.854 [2024-11-06 09:12:56.852501] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.854 [2024-11-06 09:12:56.865244] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.855 [2024-11-06 09:12:56.865259] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.855 [2024-11-06 09:12:56.880132] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.855 [2024-11-06 09:12:56.880148] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.855 [2024-11-06 09:12:56.892420] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.855 [2024-11-06 09:12:56.892436] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.855 [2024-11-06 09:12:56.904989] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.855 [2024-11-06 09:12:56.905004] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.855 [2024-11-06 09:12:56.919808] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.855 [2024-11-06 09:12:56.919824] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.855 [2024-11-06 09:12:56.932911] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.855 [2024-11-06 09:12:56.932926] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.855 [2024-11-06 09:12:56.947527] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.855 [2024-11-06 09:12:56.947543] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:06.855 [2024-11-06 09:12:56.960200] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:06.855 [2024-11-06 09:12:56.960215] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:56.972784] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:56.972800] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:56.985173] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:56.985188] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:56.999972] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:56.999987] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.012372] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.012387] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.025429] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.025444] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.039791] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.039807] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.052643] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.052659] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.064462] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.064478] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.077152] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.077167] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.092019] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.092035] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.104733] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.104753] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.120109] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.120125] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.133037] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.133053] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.147685] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.147700] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.160792] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.160808] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.175951] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.175968] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.188918] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.188934] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.203665] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.203681] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.115 [2024-11-06 09:12:57.216400] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.115 [2024-11-06 09:12:57.216416] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 18913.67 IOPS, 147.76 MiB/s [2024-11-06T08:12:57.489Z] [2024-11-06 09:12:57.228008] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.228023] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.240676] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.240692] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.252930] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.252945] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.268003] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.268018] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.281222] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.281238] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.295641] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.295657] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.308502] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.308517] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.320917] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.320932] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.335803] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.335818] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.348958] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.348973] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.363441] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.363456] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.376493] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.376512] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.389156] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.389171] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.403927] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.403942] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.416858] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.416872] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.432053] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.432068] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.444720] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.444735] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.456345] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.456360] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.469349] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.469364] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.376 [2024-11-06 09:12:57.483483] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.376 [2024-11-06 09:12:57.483498] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.496271] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.496287] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.508891] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.508905] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.524138] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.524154] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.537148] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.537162] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.551738] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.551757] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.564690] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.564704] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.579419] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.579434] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.591959] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.591976] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.604646] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.604661] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.616609] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.616625] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.628881] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.628900] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.637 [2024-11-06 09:12:57.644044] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.637 [2024-11-06 09:12:57.644059] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.638 [2024-11-06 09:12:57.656876] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.638 [2024-11-06 09:12:57.656891] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.638 [2024-11-06 09:12:57.672080] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.638 [2024-11-06 09:12:57.672095] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.638 [2024-11-06 09:12:57.685224] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.638 [2024-11-06 09:12:57.685240] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.638 [2024-11-06 09:12:57.699472] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.638 [2024-11-06 09:12:57.699487] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.638 [2024-11-06 09:12:57.712142] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.638 [2024-11-06 09:12:57.712158] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.638 [2024-11-06 09:12:57.724832] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.638 [2024-11-06 09:12:57.724847] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.638 [2024-11-06 09:12:57.739817] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.638 [2024-11-06 09:12:57.739832] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.752579] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.752595] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.764386] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.764401] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.777669] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.777684] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.792045] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.792061] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.804655] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.804669] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.819521] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.819537] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.832440] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.832456] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.844896] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.844911] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.859385] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.859401] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.871950] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.871966] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.884643] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.884662] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.896798] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.896813] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.911648] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.911664] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.924478] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.924493] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.936708] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.936723] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.948462] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.948479] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.961318] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.961333] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.976686] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.976701] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:57.991588] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:57.991604] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:07.899 [2024-11-06 09:12:58.004677] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:07.899 [2024-11-06 09:12:58.004693] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.016599] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.016615] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.029102] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.029117] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.043655] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.043670] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.056328] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.056344] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.068271] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.068286] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.080859] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.080874] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.095489] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.095505] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.108694] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.108709] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.123372] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.123389] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.136330] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.136350] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.148981] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.148996] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.163934] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.163949] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.177089] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.177105] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.192101] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.192117] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.204830] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.204845] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.219503] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.219519] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 18907.50 IOPS, 147.71 MiB/s [2024-11-06T08:12:58.273Z] [2024-11-06 09:12:58.232688] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.232703] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.247614] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.247630] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.160 [2024-11-06 09:12:58.260646] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.160 [2024-11-06 09:12:58.260662] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.422 [2024-11-06 09:12:58.273315] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.422 [2024-11-06 09:12:58.273330] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.422 [2024-11-06 09:12:58.288250] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.422 [2024-11-06 09:12:58.288265] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.422 [2024-11-06 09:12:58.301294] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.422 [2024-11-06 09:12:58.301310] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.422 [2024-11-06 09:12:58.315695] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.422 [2024-11-06 09:12:58.315711] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.422 [2024-11-06 09:12:58.328425] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.422 [2024-11-06 09:12:58.328441] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.422 [2024-11-06 09:12:58.340164] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.422 [2024-11-06 09:12:58.340180] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.422 [2024-11-06 09:12:58.353417] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.422 [2024-11-06 09:12:58.353433] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.367771] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.367787] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.380805] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.380820] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.395633] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.395649] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.408705] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.408720] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.420355] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.420371] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.433486] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.433501] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.448274] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.448290] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.461026] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.461041] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.475495] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.475510] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.488181] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.488196] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.500635] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.500650] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.513152] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.513167] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.423 [2024-11-06 09:12:58.528085] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.423 [2024-11-06 09:12:58.528100] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.540575] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.540592] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.552430] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.552446] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.565131] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.565146] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.580123] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.580139] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.592989] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.593004] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.607984] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.608000] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.620570] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.620585] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.633161] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.633176] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.647875] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.647890] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.660758] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.660773] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.676232] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.676247] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.689258] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.689273] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.703987] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.704003] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.716756] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.716771] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.731765] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.731781] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.744326] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.684 [2024-11-06 09:12:58.744342] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.684 [2024-11-06 09:12:58.756959] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.685 [2024-11-06 09:12:58.756974] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.685 [2024-11-06 09:12:58.771733] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.685 [2024-11-06 09:12:58.771753] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.685 [2024-11-06 09:12:58.784622] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.685 [2024-11-06 09:12:58.784638] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.795927] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.795944] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.808930] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.808945] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.824026] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.824042] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.837034] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.837050] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.851905] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.851921] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.865223] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.865238] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.879676] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.879691] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.892568] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.892591] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.904967] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.904982] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.919726] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.919742] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.932221] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.932236] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.944426] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.944441] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.957355] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.957369] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.971246] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.971261] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.984168] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.984184] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:58.997026] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:58.997041] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:59.011716] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:59.011732] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:59.024950] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:59.024966] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:59.039526] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:59.039541] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:08.946 [2024-11-06 09:12:59.052328] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:08.946 [2024-11-06 09:12:59.052343] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.064918] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.064933] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.079952] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.079967] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.092789] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.092804] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.104684] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.104700] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.119658] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.119673] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.132815] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.132830] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.147956] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.147974] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.160856] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.160871] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.175530] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.175545] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.188260] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.188275] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.200331] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.200346] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.213116] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.213131] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 [2024-11-06 09:12:59.227442] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.227458] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.207 18917.00 IOPS, 147.79 MiB/s 00:37:09.207 Latency(us) 00:37:09.207 [2024-11-06T08:12:59.320Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:37:09.207 Job: Nvme1n1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 128, IO size: 8192) 00:37:09.207 Nvme1n1 : 5.01 18918.49 147.80 0.00 0.00 6759.55 2635.09 11632.64 00:37:09.207 [2024-11-06T08:12:59.320Z] =================================================================================================================== 00:37:09.207 [2024-11-06T08:12:59.320Z] Total : 18918.49 147.80 0.00 0.00 6759.55 2635.09 11632.64 00:37:09.207 [2024-11-06 09:12:59.236532] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.207 [2024-11-06 09:12:59.236546] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.208 [2024-11-06 09:12:59.248530] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.208 [2024-11-06 09:12:59.248543] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.208 [2024-11-06 09:12:59.260535] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.208 [2024-11-06 09:12:59.260548] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.208 [2024-11-06 09:12:59.272533] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.208 [2024-11-06 09:12:59.272544] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.208 [2024-11-06 09:12:59.284531] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.208 [2024-11-06 09:12:59.284542] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.208 [2024-11-06 09:12:59.296528] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.208 [2024-11-06 09:12:59.296538] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.208 [2024-11-06 09:12:59.308527] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.208 [2024-11-06 09:12:59.308536] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.469 [2024-11-06 09:12:59.320529] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.469 [2024-11-06 09:12:59.320540] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.469 [2024-11-06 09:12:59.332528] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.469 [2024-11-06 09:12:59.332537] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.469 [2024-11-06 09:12:59.344527] subsystem.c:2124:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:37:09.469 [2024-11-06 09:12:59.344542] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:09.469 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh: line 42: kill: (2150130) - No such process 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@49 -- # wait 2150130 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@52 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@53 -- # rpc_cmd bdev_delay_create -b malloc0 -d delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:37:09.469 delay0 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@54 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 delay0 -n 1 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:09.469 09:12:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -c 0x1 -t 5 -q 64 -w randrw -M 50 -l warning -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 ns:1' 00:37:09.469 [2024-11-06 09:12:59.493333] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:37:17.608 Initializing NVMe Controllers 00:37:17.608 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:37:17.608 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:37:17.608 Initialization complete. Launching workers. 00:37:17.608 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 I/O completed: 300, failed: 9924 00:37:17.608 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) abort submitted 10168, failed to submit 56 00:37:17.608 success 10020, unsuccessful 148, failed 0 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@59 -- # trap - SIGINT SIGTERM EXIT 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- target/zcopy.sh@60 -- # nvmftestfini 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@514 -- # nvmfcleanup 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@121 -- # sync 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@124 -- # set +e 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@125 -- # for i in {1..20} 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:37:17.608 rmmod nvme_tcp 00:37:17.608 rmmod nvme_fabrics 00:37:17.608 rmmod nvme_keyring 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@128 -- # set -e 00:37:17.608 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@129 -- # return 0 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@515 -- # '[' -n 2148027 ']' 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@516 -- # killprocess 2148027 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@950 -- # '[' -z 2148027 ']' 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@954 -- # kill -0 2148027 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@955 -- # uname 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2148027 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2148027' 00:37:17.609 killing process with pid 2148027 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@969 -- # kill 2148027 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@974 -- # wait 2148027 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@297 -- # iptr 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@789 -- # iptables-restore 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@789 -- # iptables-save 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@302 -- # remove_spdk_ns 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:37:17.609 09:13:06 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:37:18.994 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:37:18.994 00:37:18.994 real 0m33.848s 00:37:18.994 user 0m43.812s 00:37:18.994 sys 0m11.835s 00:37:18.994 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@1126 -- # xtrace_disable 00:37:18.994 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:37:18.994 ************************************ 00:37:18.994 END TEST nvmf_zcopy 00:37:18.994 ************************************ 00:37:18.994 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@33 -- # run_test nvmf_nmic /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp --interrupt-mode 00:37:18.994 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:37:18.995 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:37:18.995 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:37:18.995 ************************************ 00:37:18.995 START TEST nvmf_nmic 00:37:18.995 ************************************ 00:37:18.995 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp --interrupt-mode 00:37:18.995 * Looking for test storage... 00:37:18.995 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:37:18.995 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:37:18.995 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1689 -- # lcov --version 00:37:18.995 09:13:08 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@333 -- # local ver1 ver1_l 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@334 -- # local ver2 ver2_l 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@336 -- # IFS=.-: 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@336 -- # read -ra ver1 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@337 -- # IFS=.-: 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@337 -- # read -ra ver2 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@338 -- # local 'op=<' 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@340 -- # ver1_l=2 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@341 -- # ver2_l=1 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@344 -- # case "$op" in 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@345 -- # : 1 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@364 -- # (( v = 0 )) 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@365 -- # decimal 1 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@353 -- # local d=1 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@355 -- # echo 1 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@365 -- # ver1[v]=1 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@366 -- # decimal 2 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@353 -- # local d=2 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@355 -- # echo 2 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@366 -- # ver2[v]=2 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@368 -- # return 0 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:37:18.995 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:18.995 --rc genhtml_branch_coverage=1 00:37:18.995 --rc genhtml_function_coverage=1 00:37:18.995 --rc genhtml_legend=1 00:37:18.995 --rc geninfo_all_blocks=1 00:37:18.995 --rc geninfo_unexecuted_blocks=1 00:37:18.995 00:37:18.995 ' 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:37:18.995 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:18.995 --rc genhtml_branch_coverage=1 00:37:18.995 --rc genhtml_function_coverage=1 00:37:18.995 --rc genhtml_legend=1 00:37:18.995 --rc geninfo_all_blocks=1 00:37:18.995 --rc geninfo_unexecuted_blocks=1 00:37:18.995 00:37:18.995 ' 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:37:18.995 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:18.995 --rc genhtml_branch_coverage=1 00:37:18.995 --rc genhtml_function_coverage=1 00:37:18.995 --rc genhtml_legend=1 00:37:18.995 --rc geninfo_all_blocks=1 00:37:18.995 --rc geninfo_unexecuted_blocks=1 00:37:18.995 00:37:18.995 ' 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:37:18.995 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:18.995 --rc genhtml_branch_coverage=1 00:37:18.995 --rc genhtml_function_coverage=1 00:37:18.995 --rc genhtml_legend=1 00:37:18.995 --rc geninfo_all_blocks=1 00:37:18.995 --rc geninfo_unexecuted_blocks=1 00:37:18.995 00:37:18.995 ' 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@7 -- # uname -s 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@15 -- # shopt -s extglob 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@5 -- # export PATH 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:18.995 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@51 -- # : 0 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@55 -- # have_pci_nics=0 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@11 -- # MALLOC_BDEV_SIZE=64 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@14 -- # nvmftestinit 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@474 -- # prepare_net_devs 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@436 -- # local -g is_hw=no 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@438 -- # remove_spdk_ns 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@309 -- # xtrace_disable 00:37:18.996 09:13:09 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@315 -- # pci_devs=() 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@315 -- # local -a pci_devs 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@316 -- # pci_net_devs=() 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@317 -- # pci_drivers=() 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@317 -- # local -A pci_drivers 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@319 -- # net_devs=() 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@319 -- # local -ga net_devs 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@320 -- # e810=() 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@320 -- # local -ga e810 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@321 -- # x722=() 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@321 -- # local -ga x722 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@322 -- # mlx=() 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@322 -- # local -ga mlx 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:37:27.144 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:37:27.145 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:37:27.145 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@416 -- # [[ up == up ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:37:27.145 Found net devices under 0000:4b:00.0: cvl_0_0 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@416 -- # [[ up == up ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:37:27.145 Found net devices under 0000:4b:00.1: cvl_0_1 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@440 -- # is_hw=yes 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:37:27.145 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:37:27.145 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.605 ms 00:37:27.145 00:37:27.145 --- 10.0.0.2 ping statistics --- 00:37:27.145 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:37:27.145 rtt min/avg/max/mdev = 0.605/0.605/0.605/0.000 ms 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:37:27.145 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:37:27.145 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.322 ms 00:37:27.145 00:37:27.145 --- 10.0.0.1 ping statistics --- 00:37:27.145 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:37:27.145 rtt min/avg/max/mdev = 0.322/0.322/0.322/0.000 ms 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@448 -- # return 0 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@15 -- # nvmfappstart -m 0xF 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@724 -- # xtrace_disable 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@507 -- # nvmfpid=2156783 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@508 -- # waitforlisten 2156783 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@831 -- # '[' -z 2156783 ']' 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:37:27.145 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@836 -- # local max_retries=100 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:37:27.146 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@840 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 [2024-11-06 09:13:16.481219] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:37:27.146 [2024-11-06 09:13:16.482164] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:37:27.146 [2024-11-06 09:13:16.482203] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:37:27.146 [2024-11-06 09:13:16.562485] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:37:27.146 [2024-11-06 09:13:16.599256] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:37:27.146 [2024-11-06 09:13:16.599291] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:37:27.146 [2024-11-06 09:13:16.599300] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:37:27.146 [2024-11-06 09:13:16.599306] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:37:27.146 [2024-11-06 09:13:16.599312] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:37:27.146 [2024-11-06 09:13:16.600784] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:37:27.146 [2024-11-06 09:13:16.600848] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:37:27.146 [2024-11-06 09:13:16.601169] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:37:27.146 [2024-11-06 09:13:16.601168] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:37:27.146 [2024-11-06 09:13:16.657140] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:37:27.146 [2024-11-06 09:13:16.657447] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:37:27.146 [2024-11-06 09:13:16.658370] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:37:27.146 [2024-11-06 09:13:16.658634] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:37:27.146 [2024-11-06 09:13:16.658815] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@864 -- # return 0 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@730 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 [2024-11-06 09:13:16.737736] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 Malloc0 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@21 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 [2024-11-06 09:13:16.809896] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@25 -- # echo 'test case1: single bdev can'\''t be used in multiple subsystems' 00:37:27.146 test case1: single bdev can't be used in multiple subsystems 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@26 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@28 -- # nmic_status=0 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc0 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 [2024-11-06 09:13:16.845641] bdev.c:8456:bdev_open: *ERROR*: bdev Malloc0 already claimed: type exclusive_write by module NVMe-oF Target 00:37:27.146 [2024-11-06 09:13:16.845660] subsystem.c:2151:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2: bdev Malloc0 cannot be opened, error=-1 00:37:27.146 [2024-11-06 09:13:16.845668] nvmf_rpc.c:1517:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:37:27.146 request: 00:37:27.146 { 00:37:27.146 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:37:27.146 "namespace": { 00:37:27.146 "bdev_name": "Malloc0", 00:37:27.146 "no_auto_visible": false 00:37:27.146 }, 00:37:27.146 "method": "nvmf_subsystem_add_ns", 00:37:27.146 "req_id": 1 00:37:27.146 } 00:37:27.146 Got JSON-RPC error response 00:37:27.146 response: 00:37:27.146 { 00:37:27.146 "code": -32602, 00:37:27.146 "message": "Invalid parameters" 00:37:27.146 } 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@29 -- # nmic_status=1 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@31 -- # '[' 1 -eq 0 ']' 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@36 -- # echo ' Adding namespace failed - expected result.' 00:37:27.146 Adding namespace failed - expected result. 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@39 -- # echo 'test case2: host connect to nvmf target in multiple paths' 00:37:27.146 test case2: host connect to nvmf target in multiple paths 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@561 -- # xtrace_disable 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:27.146 [2024-11-06 09:13:16.857758] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:37:27.146 09:13:16 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@41 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:37:27.408 09:13:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@42 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4421 00:37:27.669 09:13:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@44 -- # waitforserial SPDKISFASTANDAWESOME 00:37:27.669 09:13:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1198 -- # local i=0 00:37:27.669 09:13:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:37:27.669 09:13:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:37:27.669 09:13:17 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1205 -- # sleep 2 00:37:29.584 09:13:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:37:29.584 09:13:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:37:29.584 09:13:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:37:29.584 09:13:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:37:29.584 09:13:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:37:29.584 09:13:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1208 -- # return 0 00:37:29.584 09:13:19 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:37:29.864 [global] 00:37:29.864 thread=1 00:37:29.864 invalidate=1 00:37:29.864 rw=write 00:37:29.864 time_based=1 00:37:29.864 runtime=1 00:37:29.864 ioengine=libaio 00:37:29.864 direct=1 00:37:29.864 bs=4096 00:37:29.864 iodepth=1 00:37:29.864 norandommap=0 00:37:29.864 numjobs=1 00:37:29.864 00:37:29.864 verify_dump=1 00:37:29.864 verify_backlog=512 00:37:29.864 verify_state_save=0 00:37:29.864 do_verify=1 00:37:29.864 verify=crc32c-intel 00:37:29.864 [job0] 00:37:29.864 filename=/dev/nvme0n1 00:37:29.864 Could not set queue depth (nvme0n1) 00:37:30.127 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:30.127 fio-3.35 00:37:30.127 Starting 1 thread 00:37:31.509 00:37:31.509 job0: (groupid=0, jobs=1): err= 0: pid=2157661: Wed Nov 6 09:13:21 2024 00:37:31.509 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:37:31.509 slat (nsec): min=27321, max=42432, avg=27785.60, stdev=719.39 00:37:31.509 clat (usec): min=753, max=1166, avg=966.95, stdev=57.92 00:37:31.509 lat (usec): min=781, max=1194, avg=994.74, stdev=57.75 00:37:31.509 clat percentiles (usec): 00:37:31.509 | 1.00th=[ 783], 5.00th=[ 865], 10.00th=[ 889], 20.00th=[ 930], 00:37:31.509 | 30.00th=[ 955], 40.00th=[ 963], 50.00th=[ 971], 60.00th=[ 979], 00:37:31.509 | 70.00th=[ 996], 80.00th=[ 1012], 90.00th=[ 1037], 95.00th=[ 1057], 00:37:31.509 | 99.00th=[ 1106], 99.50th=[ 1106], 99.90th=[ 1172], 99.95th=[ 1172], 00:37:31.509 | 99.99th=[ 1172] 00:37:31.509 write: IOPS=723, BW=2893KiB/s (2963kB/s)(2896KiB/1001msec); 0 zone resets 00:37:31.509 slat (usec): min=9, max=29912, avg=72.53, stdev=1110.60 00:37:31.509 clat (usec): min=287, max=1901, avg=592.37, stdev=112.81 00:37:31.509 lat (usec): min=297, max=30642, avg=664.90, stdev=1121.85 00:37:31.509 clat percentiles (usec): 00:37:31.509 | 1.00th=[ 338], 5.00th=[ 404], 10.00th=[ 449], 20.00th=[ 510], 00:37:31.509 | 30.00th=[ 545], 40.00th=[ 578], 50.00th=[ 594], 60.00th=[ 619], 00:37:31.509 | 70.00th=[ 652], 80.00th=[ 685], 90.00th=[ 717], 95.00th=[ 742], 00:37:31.509 | 99.00th=[ 791], 99.50th=[ 816], 99.90th=[ 1909], 99.95th=[ 1909], 00:37:31.509 | 99.99th=[ 1909] 00:37:31.509 bw ( KiB/s): min= 4096, max= 4096, per=100.00%, avg=4096.00, stdev= 0.00, samples=1 00:37:31.509 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:37:31.509 lat (usec) : 500=10.92%, 750=45.47%, 1000=32.44% 00:37:31.509 lat (msec) : 2=11.17% 00:37:31.509 cpu : usr=3.00%, sys=4.40%, ctx=1241, majf=0, minf=1 00:37:31.509 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:31.509 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:31.509 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:31.509 issued rwts: total=512,724,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:31.509 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:31.509 00:37:31.509 Run status group 0 (all jobs): 00:37:31.509 READ: bw=2046KiB/s (2095kB/s), 2046KiB/s-2046KiB/s (2095kB/s-2095kB/s), io=2048KiB (2097kB), run=1001-1001msec 00:37:31.509 WRITE: bw=2893KiB/s (2963kB/s), 2893KiB/s-2893KiB/s (2963kB/s-2963kB/s), io=2896KiB (2966kB), run=1001-1001msec 00:37:31.509 00:37:31.509 Disk stats (read/write): 00:37:31.509 nvme0n1: ios=537/563, merge=0/0, ticks=1482/272, in_queue=1754, util=98.80% 00:37:31.509 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@48 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:37:31.509 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 2 controller(s) 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@49 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1219 -- # local i=0 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1231 -- # return 0 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- target/nmic.sh@53 -- # nvmftestfini 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@514 -- # nvmfcleanup 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@121 -- # sync 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@124 -- # set +e 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@125 -- # for i in {1..20} 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:37:31.510 rmmod nvme_tcp 00:37:31.510 rmmod nvme_fabrics 00:37:31.510 rmmod nvme_keyring 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@128 -- # set -e 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@129 -- # return 0 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@515 -- # '[' -n 2156783 ']' 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@516 -- # killprocess 2156783 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@950 -- # '[' -z 2156783 ']' 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@954 -- # kill -0 2156783 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@955 -- # uname 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2156783 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2156783' 00:37:31.510 killing process with pid 2156783 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@969 -- # kill 2156783 00:37:31.510 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@974 -- # wait 2156783 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@297 -- # iptr 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@789 -- # iptables-save 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@789 -- # iptables-restore 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@302 -- # remove_spdk_ns 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:37:31.771 09:13:21 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:37:33.683 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:37:33.683 00:37:33.683 real 0m14.949s 00:37:33.683 user 0m38.073s 00:37:33.683 sys 0m7.342s 00:37:33.683 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@1126 -- # xtrace_disable 00:37:33.683 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:37:33.683 ************************************ 00:37:33.683 END TEST nvmf_nmic 00:37:33.683 ************************************ 00:37:33.943 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@34 -- # run_test nvmf_fio_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp --interrupt-mode 00:37:33.943 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:37:33.943 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:37:33.943 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:37:33.943 ************************************ 00:37:33.943 START TEST nvmf_fio_target 00:37:33.943 ************************************ 00:37:33.943 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp --interrupt-mode 00:37:33.943 * Looking for test storage... 00:37:33.943 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:37:33.943 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:37:33.943 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1689 -- # lcov --version 00:37:33.943 09:13:23 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@333 -- # local ver1 ver1_l 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@334 -- # local ver2 ver2_l 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@336 -- # IFS=.-: 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@336 -- # read -ra ver1 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@337 -- # IFS=.-: 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@337 -- # read -ra ver2 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@338 -- # local 'op=<' 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@340 -- # ver1_l=2 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@341 -- # ver2_l=1 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@344 -- # case "$op" in 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@345 -- # : 1 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@364 -- # (( v = 0 )) 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@365 -- # decimal 1 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@353 -- # local d=1 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@355 -- # echo 1 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@365 -- # ver1[v]=1 00:37:33.943 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@366 -- # decimal 2 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@353 -- # local d=2 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@355 -- # echo 2 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@366 -- # ver2[v]=2 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@368 -- # return 0 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:37:34.205 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:34.205 --rc genhtml_branch_coverage=1 00:37:34.205 --rc genhtml_function_coverage=1 00:37:34.205 --rc genhtml_legend=1 00:37:34.205 --rc geninfo_all_blocks=1 00:37:34.205 --rc geninfo_unexecuted_blocks=1 00:37:34.205 00:37:34.205 ' 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:37:34.205 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:34.205 --rc genhtml_branch_coverage=1 00:37:34.205 --rc genhtml_function_coverage=1 00:37:34.205 --rc genhtml_legend=1 00:37:34.205 --rc geninfo_all_blocks=1 00:37:34.205 --rc geninfo_unexecuted_blocks=1 00:37:34.205 00:37:34.205 ' 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:37:34.205 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:34.205 --rc genhtml_branch_coverage=1 00:37:34.205 --rc genhtml_function_coverage=1 00:37:34.205 --rc genhtml_legend=1 00:37:34.205 --rc geninfo_all_blocks=1 00:37:34.205 --rc geninfo_unexecuted_blocks=1 00:37:34.205 00:37:34.205 ' 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:37:34.205 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:37:34.205 --rc genhtml_branch_coverage=1 00:37:34.205 --rc genhtml_function_coverage=1 00:37:34.205 --rc genhtml_legend=1 00:37:34.205 --rc geninfo_all_blocks=1 00:37:34.205 --rc geninfo_unexecuted_blocks=1 00:37:34.205 00:37:34.205 ' 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@7 -- # uname -s 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:37:34.205 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@15 -- # shopt -s extglob 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@5 -- # export PATH 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@51 -- # : 0 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@55 -- # have_pci_nics=0 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@16 -- # nvmftestinit 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@474 -- # prepare_net_devs 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@436 -- # local -g is_hw=no 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@438 -- # remove_spdk_ns 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@309 -- # xtrace_disable 00:37:34.206 09:13:24 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@315 -- # pci_devs=() 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@315 -- # local -a pci_devs 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@316 -- # pci_net_devs=() 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@317 -- # pci_drivers=() 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@317 -- # local -A pci_drivers 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@319 -- # net_devs=() 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@319 -- # local -ga net_devs 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@320 -- # e810=() 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@320 -- # local -ga e810 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@321 -- # x722=() 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@321 -- # local -ga x722 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@322 -- # mlx=() 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@322 -- # local -ga mlx 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:37:40.791 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:37:40.791 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:37:40.791 Found net devices under 0000:4b:00.0: cvl_0_0 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ up == up ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:37:40.791 Found net devices under 0000:4b:00.1: cvl_0_1 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@440 -- # is_hw=yes 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:37:40.791 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:37:40.792 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:37:41.053 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:37:41.053 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:37:41.053 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:37:41.053 09:13:30 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:37:41.053 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:37:41.053 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.678 ms 00:37:41.053 00:37:41.053 --- 10.0.0.2 ping statistics --- 00:37:41.053 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:37:41.053 rtt min/avg/max/mdev = 0.678/0.678/0.678/0.000 ms 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:37:41.053 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:37:41.053 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.284 ms 00:37:41.053 00:37:41.053 --- 10.0.0.1 ping statistics --- 00:37:41.053 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:37:41.053 rtt min/avg/max/mdev = 0.284/0.284/0.284/0.000 ms 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@448 -- # return 0 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@17 -- # nvmfappstart -m 0xF 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@724 -- # xtrace_disable 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@507 -- # nvmfpid=2161990 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@508 -- # waitforlisten 2161990 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0xF 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@831 -- # '[' -z 2161990 ']' 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@836 -- # local max_retries=100 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:37:41.053 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@840 -- # xtrace_disable 00:37:41.053 09:13:31 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:37:41.314 [2024-11-06 09:13:31.215315] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:37:41.314 [2024-11-06 09:13:31.216483] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:37:41.314 [2024-11-06 09:13:31.216537] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:37:41.314 [2024-11-06 09:13:31.298801] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:37:41.314 [2024-11-06 09:13:31.340232] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:37:41.314 [2024-11-06 09:13:31.340266] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:37:41.314 [2024-11-06 09:13:31.340274] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:37:41.314 [2024-11-06 09:13:31.340281] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:37:41.314 [2024-11-06 09:13:31.340287] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:37:41.314 [2024-11-06 09:13:31.341802] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:37:41.314 [2024-11-06 09:13:31.342076] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:37:41.314 [2024-11-06 09:13:31.342235] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:37:41.314 [2024-11-06 09:13:31.342235] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:37:41.314 [2024-11-06 09:13:31.399154] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:37:41.314 [2024-11-06 09:13:31.399474] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:37:41.314 [2024-11-06 09:13:31.400343] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:37:41.314 [2024-11-06 09:13:31.400542] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:37:41.314 [2024-11-06 09:13:31.400721] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:37:42.256 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:37:42.256 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@864 -- # return 0 00:37:42.256 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:37:42.256 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@730 -- # xtrace_disable 00:37:42.256 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:37:42.256 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:37:42.256 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:37:42.256 [2024-11-06 09:13:32.215200] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:37:42.256 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:37:42.516 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@21 -- # malloc_bdevs='Malloc0 ' 00:37:42.516 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:37:42.777 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@22 -- # malloc_bdevs+=Malloc1 00:37:42.777 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:37:42.777 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@24 -- # raid_malloc_bdevs='Malloc2 ' 00:37:42.777 09:13:32 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:37:43.037 09:13:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@25 -- # raid_malloc_bdevs+=Malloc3 00:37:43.037 09:13:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc2 Malloc3' 00:37:43.297 09:13:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:37:43.557 09:13:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@29 -- # concat_malloc_bdevs='Malloc4 ' 00:37:43.557 09:13:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:37:43.557 09:13:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@30 -- # concat_malloc_bdevs+='Malloc5 ' 00:37:43.557 09:13:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:37:43.816 09:13:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@31 -- # concat_malloc_bdevs+=Malloc6 00:37:43.816 09:13:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n concat0 -r concat -z 64 -b 'Malloc4 Malloc5 Malloc6' 00:37:44.076 09:13:33 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:37:44.076 09:13:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:37:44.076 09:13:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:37:44.337 09:13:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:37:44.337 09:13:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:37:44.596 09:13:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:37:44.596 [2024-11-06 09:13:34.614959] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:37:44.596 09:13:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 raid0 00:37:44.855 09:13:34 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 concat0 00:37:45.114 09:13:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@46 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:37:45.375 09:13:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@48 -- # waitforserial SPDKISFASTANDAWESOME 4 00:37:45.375 09:13:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1198 -- # local i=0 00:37:45.375 09:13:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:37:45.375 09:13:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1200 -- # [[ -n 4 ]] 00:37:45.375 09:13:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1201 -- # nvme_device_counter=4 00:37:45.375 09:13:35 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1205 -- # sleep 2 00:37:47.917 09:13:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:37:47.917 09:13:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:37:47.917 09:13:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:37:47.918 09:13:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1207 -- # nvme_devices=4 00:37:47.918 09:13:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:37:47.918 09:13:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1208 -- # return 0 00:37:47.918 09:13:37 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:37:47.918 [global] 00:37:47.918 thread=1 00:37:47.918 invalidate=1 00:37:47.918 rw=write 00:37:47.918 time_based=1 00:37:47.918 runtime=1 00:37:47.918 ioengine=libaio 00:37:47.918 direct=1 00:37:47.918 bs=4096 00:37:47.918 iodepth=1 00:37:47.918 norandommap=0 00:37:47.918 numjobs=1 00:37:47.918 00:37:47.918 verify_dump=1 00:37:47.918 verify_backlog=512 00:37:47.918 verify_state_save=0 00:37:47.918 do_verify=1 00:37:47.918 verify=crc32c-intel 00:37:47.918 [job0] 00:37:47.918 filename=/dev/nvme0n1 00:37:47.918 [job1] 00:37:47.918 filename=/dev/nvme0n2 00:37:47.918 [job2] 00:37:47.918 filename=/dev/nvme0n3 00:37:47.918 [job3] 00:37:47.918 filename=/dev/nvme0n4 00:37:47.918 Could not set queue depth (nvme0n1) 00:37:47.918 Could not set queue depth (nvme0n2) 00:37:47.918 Could not set queue depth (nvme0n3) 00:37:47.918 Could not set queue depth (nvme0n4) 00:37:47.918 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:47.918 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:47.918 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:47.918 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:47.918 fio-3.35 00:37:47.918 Starting 4 threads 00:37:49.301 00:37:49.301 job0: (groupid=0, jobs=1): err= 0: pid=2163442: Wed Nov 6 09:13:39 2024 00:37:49.301 read: IOPS=2513, BW=9.82MiB/s (10.3MB/s)(9.83MiB/1001msec) 00:37:49.301 slat (nsec): min=4112, max=46664, avg=6499.09, stdev=4401.93 00:37:49.301 clat (usec): min=142, max=538, avg=234.24, stdev=48.56 00:37:49.301 lat (usec): min=147, max=563, avg=240.74, stdev=51.81 00:37:49.301 clat percentiles (usec): 00:37:49.301 | 1.00th=[ 174], 5.00th=[ 186], 10.00th=[ 192], 20.00th=[ 208], 00:37:49.301 | 30.00th=[ 217], 40.00th=[ 223], 50.00th=[ 227], 60.00th=[ 233], 00:37:49.301 | 70.00th=[ 237], 80.00th=[ 245], 90.00th=[ 269], 95.00th=[ 318], 00:37:49.301 | 99.00th=[ 461], 99.50th=[ 482], 99.90th=[ 502], 99.95th=[ 510], 00:37:49.301 | 99.99th=[ 537] 00:37:49.301 write: IOPS=2557, BW=9.99MiB/s (10.5MB/s)(10.0MiB/1001msec); 0 zone resets 00:37:49.301 slat (nsec): min=5185, max=51792, avg=8592.49, stdev=6650.75 00:37:49.301 clat (usec): min=105, max=374, avg=141.73, stdev=41.39 00:37:49.301 lat (usec): min=111, max=406, avg=150.32, stdev=46.86 00:37:49.301 clat percentiles (usec): 00:37:49.301 | 1.00th=[ 114], 5.00th=[ 117], 10.00th=[ 119], 20.00th=[ 123], 00:37:49.301 | 30.00th=[ 126], 40.00th=[ 129], 50.00th=[ 133], 60.00th=[ 135], 00:37:49.301 | 70.00th=[ 137], 80.00th=[ 141], 90.00th=[ 155], 95.00th=[ 269], 00:37:49.301 | 99.00th=[ 322], 99.50th=[ 338], 99.90th=[ 371], 99.95th=[ 375], 00:37:49.301 | 99.99th=[ 375] 00:37:49.301 bw ( KiB/s): min=11008, max=11008, per=69.54%, avg=11008.00, stdev= 0.00, samples=1 00:37:49.301 iops : min= 2752, max= 2752, avg=2752.00, stdev= 0.00, samples=1 00:37:49.301 lat (usec) : 250=88.89%, 500=11.01%, 750=0.10% 00:37:49.301 cpu : usr=1.80%, sys=4.20%, ctx=5076, majf=0, minf=1 00:37:49.301 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:49.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:49.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:49.301 issued rwts: total=2516,2560,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:49.301 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:49.301 job1: (groupid=0, jobs=1): err= 0: pid=2163462: Wed Nov 6 09:13:39 2024 00:37:49.301 read: IOPS=385, BW=1542KiB/s (1579kB/s)(1596KiB/1035msec) 00:37:49.301 slat (nsec): min=6589, max=43788, avg=21547.06, stdev=7735.31 00:37:49.301 clat (usec): min=171, max=42482, avg=2153.45, stdev=8137.73 00:37:49.301 lat (usec): min=190, max=42491, avg=2175.00, stdev=8138.03 00:37:49.301 clat percentiles (usec): 00:37:49.301 | 1.00th=[ 208], 5.00th=[ 285], 10.00th=[ 306], 20.00th=[ 371], 00:37:49.301 | 30.00th=[ 449], 40.00th=[ 494], 50.00th=[ 523], 60.00th=[ 545], 00:37:49.301 | 70.00th=[ 562], 80.00th=[ 586], 90.00th=[ 644], 95.00th=[ 873], 00:37:49.301 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42730], 99.95th=[42730], 00:37:49.301 | 99.99th=[42730] 00:37:49.301 write: IOPS=494, BW=1979KiB/s (2026kB/s)(2048KiB/1035msec); 0 zone resets 00:37:49.301 slat (nsec): min=9112, max=66881, avg=24091.84, stdev=11443.41 00:37:49.301 clat (usec): min=107, max=665, avg=288.68, stdev=149.92 00:37:49.301 lat (usec): min=117, max=714, avg=312.77, stdev=157.49 00:37:49.301 clat percentiles (usec): 00:37:49.301 | 1.00th=[ 112], 5.00th=[ 115], 10.00th=[ 119], 20.00th=[ 126], 00:37:49.301 | 30.00th=[ 139], 40.00th=[ 245], 50.00th=[ 273], 60.00th=[ 310], 00:37:49.301 | 70.00th=[ 363], 80.00th=[ 408], 90.00th=[ 519], 95.00th=[ 586], 00:37:49.301 | 99.00th=[ 644], 99.50th=[ 660], 99.90th=[ 668], 99.95th=[ 668], 00:37:49.301 | 99.99th=[ 668] 00:37:49.301 bw ( KiB/s): min= 4096, max= 4096, per=25.88%, avg=4096.00, stdev= 0.00, samples=1 00:37:49.301 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:37:49.301 lat (usec) : 250=24.26%, 500=44.79%, 750=28.10%, 1000=0.99% 00:37:49.301 lat (msec) : 2=0.11%, 50=1.76% 00:37:49.301 cpu : usr=1.06%, sys=2.22%, ctx=911, majf=0, minf=1 00:37:49.301 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:49.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:49.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:49.301 issued rwts: total=399,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:49.301 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:49.301 job2: (groupid=0, jobs=1): err= 0: pid=2163479: Wed Nov 6 09:13:39 2024 00:37:49.301 read: IOPS=146, BW=586KiB/s (600kB/s)(604KiB/1031msec) 00:37:49.301 slat (nsec): min=4646, max=30122, avg=14389.36, stdev=7079.01 00:37:49.301 clat (usec): min=369, max=42883, avg=5093.20, stdev=12715.45 00:37:49.301 lat (usec): min=374, max=42910, avg=5107.59, stdev=12719.33 00:37:49.301 clat percentiles (usec): 00:37:49.301 | 1.00th=[ 379], 5.00th=[ 523], 10.00th=[ 570], 20.00th=[ 627], 00:37:49.301 | 30.00th=[ 685], 40.00th=[ 709], 50.00th=[ 734], 60.00th=[ 783], 00:37:49.301 | 70.00th=[ 840], 80.00th=[ 914], 90.00th=[41157], 95.00th=[42206], 00:37:49.301 | 99.00th=[42730], 99.50th=[42730], 99.90th=[42730], 99.95th=[42730], 00:37:49.301 | 99.99th=[42730] 00:37:49.301 write: IOPS=496, BW=1986KiB/s (2034kB/s)(2048KiB/1031msec); 0 zone resets 00:37:49.301 slat (usec): min=5, max=32835, avg=80.94, stdev=1453.52 00:37:49.301 clat (usec): min=125, max=850, avg=419.61, stdev=119.02 00:37:49.301 lat (usec): min=131, max=33300, avg=500.55, stdev=1461.44 00:37:49.301 clat percentiles (usec): 00:37:49.301 | 1.00th=[ 165], 5.00th=[ 210], 10.00th=[ 269], 20.00th=[ 314], 00:37:49.301 | 30.00th=[ 351], 40.00th=[ 396], 50.00th=[ 429], 60.00th=[ 453], 00:37:49.301 | 70.00th=[ 490], 80.00th=[ 529], 90.00th=[ 562], 95.00th=[ 594], 00:37:49.301 | 99.00th=[ 652], 99.50th=[ 742], 99.90th=[ 848], 99.95th=[ 848], 00:37:49.301 | 99.99th=[ 848] 00:37:49.301 bw ( KiB/s): min= 4096, max= 4096, per=25.88%, avg=4096.00, stdev= 0.00, samples=1 00:37:49.301 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:37:49.301 lat (usec) : 250=6.49%, 500=49.62%, 750=32.43%, 1000=8.60% 00:37:49.301 lat (msec) : 2=0.45%, 50=2.41% 00:37:49.301 cpu : usr=0.19%, sys=0.78%, ctx=668, majf=0, minf=1 00:37:49.301 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:49.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:49.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:49.301 issued rwts: total=151,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:49.301 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:49.301 job3: (groupid=0, jobs=1): err= 0: pid=2163485: Wed Nov 6 09:13:39 2024 00:37:49.301 read: IOPS=20, BW=83.5KiB/s (85.5kB/s)(84.0KiB/1006msec) 00:37:49.301 slat (nsec): min=26588, max=30303, avg=27881.86, stdev=670.03 00:37:49.301 clat (usec): min=835, max=41956, avg=37245.20, stdev=12087.52 00:37:49.301 lat (usec): min=865, max=41984, avg=37273.08, stdev=12087.33 00:37:49.301 clat percentiles (usec): 00:37:49.301 | 1.00th=[ 840], 5.00th=[ 955], 10.00th=[40633], 20.00th=[41157], 00:37:49.301 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:37:49.301 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41681], 00:37:49.301 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:37:49.301 | 99.99th=[42206] 00:37:49.301 write: IOPS=508, BW=2036KiB/s (2085kB/s)(2048KiB/1006msec); 0 zone resets 00:37:49.301 slat (nsec): min=10316, max=55824, avg=30207.37, stdev=10883.15 00:37:49.301 clat (usec): min=191, max=770, avg=397.85, stdev=106.85 00:37:49.301 lat (usec): min=226, max=782, avg=428.06, stdev=109.34 00:37:49.301 clat percentiles (usec): 00:37:49.301 | 1.00th=[ 212], 5.00th=[ 241], 10.00th=[ 285], 20.00th=[ 310], 00:37:49.302 | 30.00th=[ 326], 40.00th=[ 347], 50.00th=[ 371], 60.00th=[ 420], 00:37:49.302 | 70.00th=[ 457], 80.00th=[ 494], 90.00th=[ 562], 95.00th=[ 594], 00:37:49.302 | 99.00th=[ 644], 99.50th=[ 709], 99.90th=[ 775], 99.95th=[ 775], 00:37:49.302 | 99.99th=[ 775] 00:37:49.302 bw ( KiB/s): min= 4096, max= 4096, per=25.88%, avg=4096.00, stdev= 0.00, samples=1 00:37:49.302 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:37:49.302 lat (usec) : 250=6.00%, 500=73.17%, 750=16.70%, 1000=0.56% 00:37:49.302 lat (msec) : 50=3.56% 00:37:49.302 cpu : usr=0.90%, sys=1.39%, ctx=535, majf=0, minf=1 00:37:49.302 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:49.302 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:49.302 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:49.302 issued rwts: total=21,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:49.302 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:49.302 00:37:49.302 Run status group 0 (all jobs): 00:37:49.302 READ: bw=11.7MiB/s (12.2MB/s), 83.5KiB/s-9.82MiB/s (85.5kB/s-10.3MB/s), io=12.1MiB (12.6MB), run=1001-1035msec 00:37:49.302 WRITE: bw=15.5MiB/s (16.2MB/s), 1979KiB/s-9.99MiB/s (2026kB/s-10.5MB/s), io=16.0MiB (16.8MB), run=1001-1035msec 00:37:49.302 00:37:49.302 Disk stats (read/write): 00:37:49.302 nvme0n1: ios=2098/2097, merge=0/0, ticks=544/285, in_queue=829, util=87.47% 00:37:49.302 nvme0n2: ios=444/512, merge=0/0, ticks=740/136, in_queue=876, util=90.98% 00:37:49.302 nvme0n3: ios=224/512, merge=0/0, ticks=953/212, in_queue=1165, util=94.37% 00:37:49.302 nvme0n4: ios=73/512, merge=0/0, ticks=950/193, in_queue=1143, util=94.09% 00:37:49.302 09:13:39 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t randwrite -r 1 -v 00:37:49.302 [global] 00:37:49.302 thread=1 00:37:49.302 invalidate=1 00:37:49.302 rw=randwrite 00:37:49.302 time_based=1 00:37:49.302 runtime=1 00:37:49.302 ioengine=libaio 00:37:49.302 direct=1 00:37:49.302 bs=4096 00:37:49.302 iodepth=1 00:37:49.302 norandommap=0 00:37:49.302 numjobs=1 00:37:49.302 00:37:49.302 verify_dump=1 00:37:49.302 verify_backlog=512 00:37:49.302 verify_state_save=0 00:37:49.302 do_verify=1 00:37:49.302 verify=crc32c-intel 00:37:49.302 [job0] 00:37:49.302 filename=/dev/nvme0n1 00:37:49.302 [job1] 00:37:49.302 filename=/dev/nvme0n2 00:37:49.302 [job2] 00:37:49.302 filename=/dev/nvme0n3 00:37:49.302 [job3] 00:37:49.302 filename=/dev/nvme0n4 00:37:49.302 Could not set queue depth (nvme0n1) 00:37:49.302 Could not set queue depth (nvme0n2) 00:37:49.302 Could not set queue depth (nvme0n3) 00:37:49.302 Could not set queue depth (nvme0n4) 00:37:49.562 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:49.562 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:49.562 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:49.562 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:49.562 fio-3.35 00:37:49.562 Starting 4 threads 00:37:50.949 00:37:50.949 job0: (groupid=0, jobs=1): err= 0: pid=2163918: Wed Nov 6 09:13:40 2024 00:37:50.949 read: IOPS=17, BW=71.1KiB/s (72.9kB/s)(72.0KiB/1012msec) 00:37:50.949 slat (nsec): min=6177, max=16468, avg=9492.56, stdev=2653.33 00:37:50.949 clat (usec): min=555, max=41230, avg=38756.75, stdev=9534.32 00:37:50.949 lat (usec): min=562, max=41236, avg=38766.25, stdev=9534.94 00:37:50.949 clat percentiles (usec): 00:37:50.949 | 1.00th=[ 553], 5.00th=[ 553], 10.00th=[40633], 20.00th=[41157], 00:37:50.949 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:37:50.949 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:37:50.949 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:37:50.949 | 99.99th=[41157] 00:37:50.949 write: IOPS=505, BW=2024KiB/s (2072kB/s)(2048KiB/1012msec); 0 zone resets 00:37:50.949 slat (nsec): min=5721, max=52099, avg=25738.76, stdev=11943.29 00:37:50.949 clat (usec): min=152, max=1013, avg=580.09, stdev=149.92 00:37:50.949 lat (usec): min=159, max=1046, avg=605.83, stdev=157.13 00:37:50.949 clat percentiles (usec): 00:37:50.949 | 1.00th=[ 260], 5.00th=[ 310], 10.00th=[ 371], 20.00th=[ 437], 00:37:50.949 | 30.00th=[ 498], 40.00th=[ 553], 50.00th=[ 603], 60.00th=[ 644], 00:37:50.949 | 70.00th=[ 668], 80.00th=[ 709], 90.00th=[ 750], 95.00th=[ 799], 00:37:50.949 | 99.00th=[ 889], 99.50th=[ 955], 99.90th=[ 1012], 99.95th=[ 1012], 00:37:50.949 | 99.99th=[ 1012] 00:37:50.949 bw ( KiB/s): min= 4096, max= 4096, per=32.35%, avg=4096.00, stdev= 0.00, samples=1 00:37:50.949 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:37:50.949 lat (usec) : 250=0.75%, 500=28.49%, 750=57.55%, 1000=9.62% 00:37:50.949 lat (msec) : 2=0.38%, 50=3.21% 00:37:50.949 cpu : usr=0.69%, sys=1.29%, ctx=533, majf=0, minf=1 00:37:50.949 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:50.949 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:50.949 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:50.949 issued rwts: total=18,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:50.949 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:50.949 job1: (groupid=0, jobs=1): err= 0: pid=2163939: Wed Nov 6 09:13:40 2024 00:37:50.949 read: IOPS=512, BW=2048KiB/s (2097kB/s)(2048KiB/1000msec) 00:37:50.949 slat (nsec): min=6797, max=61070, avg=24026.73, stdev=6756.80 00:37:50.949 clat (usec): min=299, max=1306, avg=779.82, stdev=188.60 00:37:50.949 lat (usec): min=306, max=1331, avg=803.84, stdev=190.16 00:37:50.949 clat percentiles (usec): 00:37:50.949 | 1.00th=[ 375], 5.00th=[ 482], 10.00th=[ 570], 20.00th=[ 627], 00:37:50.949 | 30.00th=[ 685], 40.00th=[ 734], 50.00th=[ 775], 60.00th=[ 816], 00:37:50.949 | 70.00th=[ 848], 80.00th=[ 889], 90.00th=[ 1074], 95.00th=[ 1156], 00:37:50.949 | 99.00th=[ 1237], 99.50th=[ 1270], 99.90th=[ 1303], 99.95th=[ 1303], 00:37:50.949 | 99.99th=[ 1303] 00:37:50.949 write: IOPS=956, BW=3824KiB/s (3916kB/s)(3828KiB/1001msec); 0 zone resets 00:37:50.949 slat (nsec): min=9262, max=70572, avg=27708.40, stdev=9830.02 00:37:50.949 clat (usec): min=122, max=989, avg=575.72, stdev=179.61 00:37:50.949 lat (usec): min=134, max=1004, avg=603.43, stdev=183.10 00:37:50.949 clat percentiles (usec): 00:37:50.949 | 1.00th=[ 159], 5.00th=[ 269], 10.00th=[ 318], 20.00th=[ 412], 00:37:50.949 | 30.00th=[ 478], 40.00th=[ 537], 50.00th=[ 594], 60.00th=[ 652], 00:37:50.949 | 70.00th=[ 693], 80.00th=[ 742], 90.00th=[ 791], 95.00th=[ 840], 00:37:50.949 | 99.00th=[ 914], 99.50th=[ 938], 99.90th=[ 988], 99.95th=[ 988], 00:37:50.949 | 99.99th=[ 988] 00:37:50.949 bw ( KiB/s): min= 4096, max= 4096, per=32.35%, avg=4096.00, stdev= 0.00, samples=1 00:37:50.949 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:37:50.949 lat (usec) : 250=2.38%, 500=21.78%, 750=44.66%, 1000=26.68% 00:37:50.949 lat (msec) : 2=4.49% 00:37:50.949 cpu : usr=2.30%, sys=3.70%, ctx=1469, majf=0, minf=2 00:37:50.949 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:50.949 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:50.949 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:50.949 issued rwts: total=512,957,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:50.949 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:50.949 job2: (groupid=0, jobs=1): err= 0: pid=2163961: Wed Nov 6 09:13:40 2024 00:37:50.949 read: IOPS=511, BW=2046KiB/s (2095kB/s)(2048KiB/1001msec) 00:37:50.949 slat (nsec): min=8348, max=43490, avg=25341.88, stdev=2325.17 00:37:50.949 clat (usec): min=601, max=1240, avg=1026.48, stdev=83.09 00:37:50.949 lat (usec): min=627, max=1265, avg=1051.83, stdev=83.08 00:37:50.949 clat percentiles (usec): 00:37:50.949 | 1.00th=[ 807], 5.00th=[ 873], 10.00th=[ 922], 20.00th=[ 963], 00:37:50.949 | 30.00th=[ 996], 40.00th=[ 1012], 50.00th=[ 1029], 60.00th=[ 1057], 00:37:50.949 | 70.00th=[ 1074], 80.00th=[ 1090], 90.00th=[ 1123], 95.00th=[ 1156], 00:37:50.949 | 99.00th=[ 1205], 99.50th=[ 1237], 99.90th=[ 1237], 99.95th=[ 1237], 00:37:50.949 | 99.99th=[ 1237] 00:37:50.949 write: IOPS=709, BW=2837KiB/s (2905kB/s)(2840KiB/1001msec); 0 zone resets 00:37:50.949 slat (nsec): min=9317, max=60258, avg=27924.46, stdev=8716.35 00:37:50.949 clat (usec): min=141, max=1065, avg=609.10, stdev=132.28 00:37:50.949 lat (usec): min=160, max=1095, avg=637.03, stdev=135.57 00:37:50.949 clat percentiles (usec): 00:37:50.949 | 1.00th=[ 293], 5.00th=[ 363], 10.00th=[ 420], 20.00th=[ 498], 00:37:50.949 | 30.00th=[ 562], 40.00th=[ 594], 50.00th=[ 627], 60.00th=[ 660], 00:37:50.949 | 70.00th=[ 685], 80.00th=[ 717], 90.00th=[ 766], 95.00th=[ 799], 00:37:50.949 | 99.00th=[ 873], 99.50th=[ 914], 99.90th=[ 1074], 99.95th=[ 1074], 00:37:50.949 | 99.99th=[ 1074] 00:37:50.949 bw ( KiB/s): min= 4096, max= 4096, per=32.35%, avg=4096.00, stdev= 0.00, samples=1 00:37:50.949 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:37:50.949 lat (usec) : 250=0.33%, 500=11.70%, 750=39.12%, 1000=20.13% 00:37:50.949 lat (msec) : 2=28.72% 00:37:50.949 cpu : usr=1.80%, sys=3.40%, ctx=1222, majf=0, minf=1 00:37:50.949 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:50.949 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:50.949 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:50.949 issued rwts: total=512,710,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:50.949 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:50.949 job3: (groupid=0, jobs=1): err= 0: pid=2163972: Wed Nov 6 09:13:40 2024 00:37:50.949 read: IOPS=585, BW=2342KiB/s (2398kB/s)(2344KiB/1001msec) 00:37:50.949 slat (nsec): min=6063, max=67030, avg=25368.81, stdev=8488.17 00:37:50.949 clat (usec): min=369, max=926, avg=694.95, stdev=104.47 00:37:50.949 lat (usec): min=401, max=953, avg=720.32, stdev=105.61 00:37:50.949 clat percentiles (usec): 00:37:50.949 | 1.00th=[ 457], 5.00th=[ 529], 10.00th=[ 562], 20.00th=[ 594], 00:37:50.949 | 30.00th=[ 635], 40.00th=[ 660], 50.00th=[ 693], 60.00th=[ 725], 00:37:50.949 | 70.00th=[ 766], 80.00th=[ 799], 90.00th=[ 832], 95.00th=[ 857], 00:37:50.949 | 99.00th=[ 898], 99.50th=[ 914], 99.90th=[ 930], 99.95th=[ 930], 00:37:50.949 | 99.99th=[ 930] 00:37:50.949 write: IOPS=1022, BW=4092KiB/s (4190kB/s)(4096KiB/1001msec); 0 zone resets 00:37:50.949 slat (nsec): min=9038, max=67429, avg=31542.04, stdev=8977.44 00:37:50.949 clat (usec): min=159, max=3191, avg=520.70, stdev=187.69 00:37:50.949 lat (usec): min=169, max=3232, avg=552.24, stdev=190.07 00:37:50.949 clat percentiles (usec): 00:37:50.949 | 1.00th=[ 210], 5.00th=[ 273], 10.00th=[ 322], 20.00th=[ 371], 00:37:50.949 | 30.00th=[ 416], 40.00th=[ 474], 50.00th=[ 506], 60.00th=[ 545], 00:37:50.949 | 70.00th=[ 594], 80.00th=[ 652], 90.00th=[ 734], 95.00th=[ 816], 00:37:50.949 | 99.00th=[ 906], 99.50th=[ 955], 99.90th=[ 2311], 99.95th=[ 3195], 00:37:50.949 | 99.99th=[ 3195] 00:37:50.949 bw ( KiB/s): min= 4096, max= 4096, per=32.35%, avg=4096.00, stdev= 0.00, samples=1 00:37:50.949 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:37:50.949 lat (usec) : 250=1.49%, 500=30.50%, 750=50.25%, 1000=17.52% 00:37:50.949 lat (msec) : 2=0.12%, 4=0.12% 00:37:50.949 cpu : usr=3.40%, sys=6.10%, ctx=1611, majf=0, minf=2 00:37:50.949 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:50.949 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:50.949 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:50.949 issued rwts: total=586,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:50.949 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:50.949 00:37:50.949 Run status group 0 (all jobs): 00:37:50.949 READ: bw=6435KiB/s (6589kB/s), 71.1KiB/s-2342KiB/s (72.9kB/s-2398kB/s), io=6512KiB (6668kB), run=1000-1012msec 00:37:50.949 WRITE: bw=12.4MiB/s (13.0MB/s), 2024KiB/s-4092KiB/s (2072kB/s-4190kB/s), io=12.5MiB (13.1MB), run=1001-1012msec 00:37:50.949 00:37:50.949 Disk stats (read/write): 00:37:50.949 nvme0n1: ios=61/512, merge=0/0, ticks=642/290, in_queue=932, util=87.17% 00:37:50.949 nvme0n2: ios=562/651, merge=0/0, ticks=535/310, in_queue=845, util=89.50% 00:37:50.949 nvme0n3: ios=476/512, merge=0/0, ticks=479/294, in_queue=773, util=89.52% 00:37:50.949 nvme0n4: ios=569/713, merge=0/0, ticks=673/305, in_queue=978, util=97.94% 00:37:50.949 09:13:40 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t write -r 1 -v 00:37:50.949 [global] 00:37:50.949 thread=1 00:37:50.949 invalidate=1 00:37:50.949 rw=write 00:37:50.949 time_based=1 00:37:50.949 runtime=1 00:37:50.949 ioengine=libaio 00:37:50.949 direct=1 00:37:50.949 bs=4096 00:37:50.949 iodepth=128 00:37:50.949 norandommap=0 00:37:50.949 numjobs=1 00:37:50.949 00:37:50.949 verify_dump=1 00:37:50.949 verify_backlog=512 00:37:50.950 verify_state_save=0 00:37:50.950 do_verify=1 00:37:50.950 verify=crc32c-intel 00:37:50.950 [job0] 00:37:50.950 filename=/dev/nvme0n1 00:37:50.950 [job1] 00:37:50.950 filename=/dev/nvme0n2 00:37:50.950 [job2] 00:37:50.950 filename=/dev/nvme0n3 00:37:50.950 [job3] 00:37:50.950 filename=/dev/nvme0n4 00:37:50.950 Could not set queue depth (nvme0n1) 00:37:50.950 Could not set queue depth (nvme0n2) 00:37:50.950 Could not set queue depth (nvme0n3) 00:37:50.950 Could not set queue depth (nvme0n4) 00:37:51.211 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:37:51.211 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:37:51.211 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:37:51.211 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:37:51.211 fio-3.35 00:37:51.211 Starting 4 threads 00:37:52.597 00:37:52.597 job0: (groupid=0, jobs=1): err= 0: pid=2164371: Wed Nov 6 09:13:42 2024 00:37:52.597 read: IOPS=5701, BW=22.3MiB/s (23.4MB/s)(22.4MiB/1004msec) 00:37:52.597 slat (nsec): min=940, max=9396.9k, avg=80467.92, stdev=590692.96 00:37:52.597 clat (usec): min=2830, max=30103, avg=11814.10, stdev=5940.26 00:37:52.597 lat (usec): min=3083, max=34430, avg=11894.57, stdev=5989.04 00:37:52.597 clat percentiles (usec): 00:37:52.597 | 1.00th=[ 3556], 5.00th=[ 4490], 10.00th=[ 5407], 20.00th=[ 6456], 00:37:52.597 | 30.00th=[ 7046], 40.00th=[ 8717], 50.00th=[10159], 60.00th=[11994], 00:37:52.597 | 70.00th=[14746], 80.00th=[17695], 90.00th=[20579], 95.00th=[22938], 00:37:52.597 | 99.00th=[25560], 99.50th=[27657], 99.90th=[29230], 99.95th=[29754], 00:37:52.597 | 99.99th=[30016] 00:37:52.597 write: IOPS=6119, BW=23.9MiB/s (25.1MB/s)(24.0MiB/1004msec); 0 zone resets 00:37:52.597 slat (nsec): min=1644, max=10075k, avg=67706.38, stdev=525756.76 00:37:52.597 clat (usec): min=734, max=33572, avg=9638.13, stdev=5080.01 00:37:52.597 lat (usec): min=742, max=33581, avg=9705.83, stdev=5127.86 00:37:52.597 clat percentiles (usec): 00:37:52.597 | 1.00th=[ 3490], 5.00th=[ 4424], 10.00th=[ 4883], 20.00th=[ 5604], 00:37:52.597 | 30.00th=[ 6587], 40.00th=[ 7111], 50.00th=[ 7635], 60.00th=[ 8717], 00:37:52.597 | 70.00th=[11338], 80.00th=[13566], 90.00th=[16057], 95.00th=[20055], 00:37:52.597 | 99.00th=[28443], 99.50th=[30016], 99.90th=[32637], 99.95th=[33817], 00:37:52.597 | 99.99th=[33817] 00:37:52.597 bw ( KiB/s): min=24296, max=24576, per=29.36%, avg=24436.00, stdev=197.99, samples=2 00:37:52.597 iops : min= 6074, max= 6144, avg=6109.00, stdev=49.50, samples=2 00:37:52.597 lat (usec) : 750=0.02%, 1000=0.02% 00:37:52.597 lat (msec) : 2=0.06%, 4=2.15%, 10=55.28%, 20=33.74%, 50=8.74% 00:37:52.597 cpu : usr=4.79%, sys=6.68%, ctx=404, majf=0, minf=1 00:37:52.597 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.5% 00:37:52.597 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:52.597 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:37:52.597 issued rwts: total=5724,6144,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:52.597 latency : target=0, window=0, percentile=100.00%, depth=128 00:37:52.597 job1: (groupid=0, jobs=1): err= 0: pid=2164392: Wed Nov 6 09:13:42 2024 00:37:52.597 read: IOPS=7130, BW=27.9MiB/s (29.2MB/s)(28.0MiB/1006msec) 00:37:52.597 slat (nsec): min=1274, max=8970.4k, avg=67429.71, stdev=489916.53 00:37:52.597 clat (usec): min=3566, max=26740, avg=9087.61, stdev=2999.54 00:37:52.597 lat (usec): min=3831, max=27356, avg=9155.04, stdev=3026.34 00:37:52.597 clat percentiles (usec): 00:37:52.597 | 1.00th=[ 4555], 5.00th=[ 5538], 10.00th=[ 6259], 20.00th=[ 6849], 00:37:52.597 | 30.00th=[ 7308], 40.00th=[ 7898], 50.00th=[ 8356], 60.00th=[ 8979], 00:37:52.597 | 70.00th=[ 9765], 80.00th=[11076], 90.00th=[13042], 95.00th=[14091], 00:37:52.597 | 99.00th=[21103], 99.50th=[22152], 99.90th=[22938], 99.95th=[22938], 00:37:52.597 | 99.99th=[26870] 00:37:52.597 write: IOPS=7634, BW=29.8MiB/s (31.3MB/s)(30.0MiB/1006msec); 0 zone resets 00:37:52.597 slat (nsec): min=1981, max=6256.9k, avg=63559.70, stdev=443597.09 00:37:52.597 clat (usec): min=2771, max=29394, avg=8090.56, stdev=3251.76 00:37:52.597 lat (usec): min=2861, max=29397, avg=8154.12, stdev=3263.14 00:37:52.597 clat percentiles (usec): 00:37:52.597 | 1.00th=[ 4424], 5.00th=[ 5080], 10.00th=[ 5342], 20.00th=[ 5735], 00:37:52.597 | 30.00th=[ 6390], 40.00th=[ 7046], 50.00th=[ 7504], 60.00th=[ 8094], 00:37:52.597 | 70.00th=[ 8586], 80.00th=[ 9634], 90.00th=[10421], 95.00th=[11994], 00:37:52.597 | 99.00th=[22938], 99.50th=[25560], 99.90th=[28967], 99.95th=[29492], 00:37:52.597 | 99.99th=[29492] 00:37:52.597 bw ( KiB/s): min=27696, max=32768, per=36.33%, avg=30232.00, stdev=3586.45, samples=2 00:37:52.597 iops : min= 6924, max= 8192, avg=7558.00, stdev=896.61, samples=2 00:37:52.597 lat (msec) : 4=0.48%, 10=79.10%, 20=18.83%, 50=1.59% 00:37:52.597 cpu : usr=2.89%, sys=4.98%, ctx=446, majf=0, minf=1 00:37:52.597 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.2%, >=64=99.6% 00:37:52.597 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:52.597 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:37:52.597 issued rwts: total=7173,7680,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:52.597 latency : target=0, window=0, percentile=100.00%, depth=128 00:37:52.597 job2: (groupid=0, jobs=1): err= 0: pid=2164419: Wed Nov 6 09:13:42 2024 00:37:52.597 read: IOPS=2113, BW=8456KiB/s (8659kB/s)(8532KiB/1009msec) 00:37:52.597 slat (nsec): min=1085, max=14269k, avg=186480.34, stdev=1151389.80 00:37:52.597 clat (usec): min=6147, max=94706, avg=20630.29, stdev=14908.61 00:37:52.597 lat (usec): min=6153, max=94713, avg=20816.77, stdev=15045.66 00:37:52.597 clat percentiles (usec): 00:37:52.597 | 1.00th=[ 6259], 5.00th=[ 8455], 10.00th=[10421], 20.00th=[11731], 00:37:52.597 | 30.00th=[14353], 40.00th=[14877], 50.00th=[16909], 60.00th=[17957], 00:37:52.597 | 70.00th=[19792], 80.00th=[22676], 90.00th=[33162], 95.00th=[58459], 00:37:52.598 | 99.00th=[84411], 99.50th=[93848], 99.90th=[94897], 99.95th=[94897], 00:37:52.598 | 99.99th=[94897] 00:37:52.598 write: IOPS=2537, BW=9.91MiB/s (10.4MB/s)(10.0MiB/1009msec); 0 zone resets 00:37:52.598 slat (nsec): min=1747, max=25754k, avg=230155.54, stdev=1241109.30 00:37:52.598 clat (msec): min=5, max=115, avg=32.17, stdev=29.90 00:37:52.598 lat (msec): min=5, max=115, avg=32.40, stdev=30.10 00:37:52.598 clat percentiles (msec): 00:37:52.598 | 1.00th=[ 7], 5.00th=[ 7], 10.00th=[ 10], 20.00th=[ 13], 00:37:52.598 | 30.00th=[ 15], 40.00th=[ 18], 50.00th=[ 21], 60.00th=[ 22], 00:37:52.598 | 70.00th=[ 26], 80.00th=[ 51], 90.00th=[ 91], 95.00th=[ 101], 00:37:52.598 | 99.00th=[ 112], 99.50th=[ 115], 99.90th=[ 116], 99.95th=[ 116], 00:37:52.598 | 99.99th=[ 116] 00:37:52.598 bw ( KiB/s): min= 7464, max=12680, per=12.10%, avg=10072.00, stdev=3688.27, samples=2 00:37:52.598 iops : min= 1866, max= 3170, avg=2518.00, stdev=922.07, samples=2 00:37:52.598 lat (msec) : 10=8.91%, 20=49.73%, 50=27.42%, 100=11.06%, 250=2.88% 00:37:52.598 cpu : usr=1.98%, sys=2.58%, ctx=255, majf=0, minf=2 00:37:52.598 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.3%, 32=0.7%, >=64=98.7% 00:37:52.598 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:52.598 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:37:52.598 issued rwts: total=2133,2560,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:52.598 latency : target=0, window=0, percentile=100.00%, depth=128 00:37:52.598 job3: (groupid=0, jobs=1): err= 0: pid=2164430: Wed Nov 6 09:13:42 2024 00:37:52.598 read: IOPS=4426, BW=17.3MiB/s (18.1MB/s)(17.3MiB/1003msec) 00:37:52.598 slat (nsec): min=926, max=16409k, avg=99221.96, stdev=823113.01 00:37:52.598 clat (usec): min=1568, max=42048, avg=14114.18, stdev=6338.26 00:37:52.598 lat (usec): min=1576, max=42055, avg=14213.40, stdev=6387.85 00:37:52.598 clat percentiles (usec): 00:37:52.598 | 1.00th=[ 2245], 5.00th=[ 5276], 10.00th=[ 7898], 20.00th=[ 9372], 00:37:52.598 | 30.00th=[10028], 40.00th=[11338], 50.00th=[12649], 60.00th=[13960], 00:37:52.598 | 70.00th=[16450], 80.00th=[19792], 90.00th=[22676], 95.00th=[26870], 00:37:52.598 | 99.00th=[31327], 99.50th=[31851], 99.90th=[34866], 99.95th=[34866], 00:37:52.598 | 99.99th=[42206] 00:37:52.598 write: IOPS=4594, BW=17.9MiB/s (18.8MB/s)(18.0MiB/1003msec); 0 zone resets 00:37:52.598 slat (nsec): min=1748, max=12171k, avg=102787.58, stdev=747021.66 00:37:52.598 clat (usec): min=1101, max=88425, avg=14035.38, stdev=11306.73 00:37:52.598 lat (usec): min=1139, max=88435, avg=14138.16, stdev=11373.42 00:37:52.598 clat percentiles (usec): 00:37:52.598 | 1.00th=[ 2868], 5.00th=[ 6325], 10.00th=[ 6849], 20.00th=[ 8029], 00:37:52.598 | 30.00th=[ 9110], 40.00th=[ 9896], 50.00th=[11207], 60.00th=[12780], 00:37:52.598 | 70.00th=[13960], 80.00th=[16319], 90.00th=[21365], 95.00th=[28443], 00:37:52.598 | 99.00th=[76022], 99.50th=[81265], 99.90th=[88605], 99.95th=[88605], 00:37:52.598 | 99.99th=[88605] 00:37:52.598 bw ( KiB/s): min=16384, max=20480, per=22.15%, avg=18432.00, stdev=2896.31, samples=2 00:37:52.598 iops : min= 4096, max= 5120, avg=4608.00, stdev=724.08, samples=2 00:37:52.598 lat (msec) : 2=0.38%, 4=2.40%, 10=33.93%, 20=47.54%, 50=14.36% 00:37:52.598 lat (msec) : 100=1.40% 00:37:52.598 cpu : usr=3.29%, sys=5.09%, ctx=273, majf=0, minf=1 00:37:52.598 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:37:52.598 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:52.598 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:37:52.598 issued rwts: total=4440,4608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:52.598 latency : target=0, window=0, percentile=100.00%, depth=128 00:37:52.598 00:37:52.598 Run status group 0 (all jobs): 00:37:52.598 READ: bw=75.4MiB/s (79.0MB/s), 8456KiB/s-27.9MiB/s (8659kB/s-29.2MB/s), io=76.1MiB (79.7MB), run=1003-1009msec 00:37:52.598 WRITE: bw=81.3MiB/s (85.2MB/s), 9.91MiB/s-29.8MiB/s (10.4MB/s-31.3MB/s), io=82.0MiB (86.0MB), run=1003-1009msec 00:37:52.598 00:37:52.598 Disk stats (read/write): 00:37:52.598 nvme0n1: ios=4342/4608, merge=0/0, ticks=34079/30404, in_queue=64483, util=86.87% 00:37:52.598 nvme0n2: ios=6194/6444, merge=0/0, ticks=50330/44159, in_queue=94489, util=86.38% 00:37:52.598 nvme0n3: ios=1586/1943, merge=0/0, ticks=29550/63158, in_queue=92708, util=93.30% 00:37:52.598 nvme0n4: ios=3442/3584, merge=0/0, ticks=46384/49552, in_queue=95936, util=95.51% 00:37:52.598 09:13:42 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t randwrite -r 1 -v 00:37:52.598 [global] 00:37:52.598 thread=1 00:37:52.598 invalidate=1 00:37:52.598 rw=randwrite 00:37:52.598 time_based=1 00:37:52.598 runtime=1 00:37:52.598 ioengine=libaio 00:37:52.598 direct=1 00:37:52.598 bs=4096 00:37:52.598 iodepth=128 00:37:52.598 norandommap=0 00:37:52.598 numjobs=1 00:37:52.598 00:37:52.598 verify_dump=1 00:37:52.598 verify_backlog=512 00:37:52.598 verify_state_save=0 00:37:52.598 do_verify=1 00:37:52.598 verify=crc32c-intel 00:37:52.598 [job0] 00:37:52.598 filename=/dev/nvme0n1 00:37:52.598 [job1] 00:37:52.598 filename=/dev/nvme0n2 00:37:52.598 [job2] 00:37:52.598 filename=/dev/nvme0n3 00:37:52.598 [job3] 00:37:52.598 filename=/dev/nvme0n4 00:37:52.598 Could not set queue depth (nvme0n1) 00:37:52.598 Could not set queue depth (nvme0n2) 00:37:52.598 Could not set queue depth (nvme0n3) 00:37:52.598 Could not set queue depth (nvme0n4) 00:37:53.167 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:37:53.167 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:37:53.167 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:37:53.167 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:37:53.167 fio-3.35 00:37:53.167 Starting 4 threads 00:37:54.112 00:37:54.112 job0: (groupid=0, jobs=1): err= 0: pid=2164843: Wed Nov 6 09:13:44 2024 00:37:54.112 read: IOPS=2015, BW=8063KiB/s (8257kB/s)(8192KiB/1016msec) 00:37:54.112 slat (nsec): min=955, max=25523k, avg=175041.50, stdev=1193302.25 00:37:54.112 clat (usec): min=11548, max=64793, avg=23091.21, stdev=9028.62 00:37:54.112 lat (usec): min=11552, max=64797, avg=23266.26, stdev=9069.27 00:37:54.112 clat percentiles (usec): 00:37:54.112 | 1.00th=[12256], 5.00th=[12911], 10.00th=[14484], 20.00th=[15533], 00:37:54.112 | 30.00th=[17957], 40.00th=[19268], 50.00th=[21365], 60.00th=[24249], 00:37:54.112 | 70.00th=[25822], 80.00th=[27132], 90.00th=[28705], 95.00th=[50594], 00:37:54.112 | 99.00th=[52691], 99.50th=[55837], 99.90th=[64750], 99.95th=[64750], 00:37:54.112 | 99.99th=[64750] 00:37:54.112 write: IOPS=2418, BW=9673KiB/s (9905kB/s)(9828KiB/1016msec); 0 zone resets 00:37:54.112 slat (nsec): min=1652, max=75945k, avg=249502.11, stdev=2276862.59 00:37:54.112 clat (msec): min=5, max=190, avg=33.06, stdev=31.29 00:37:54.112 lat (msec): min=5, max=190, avg=33.31, stdev=31.43 00:37:54.112 clat percentiles (msec): 00:37:54.112 | 1.00th=[ 8], 5.00th=[ 13], 10.00th=[ 13], 20.00th=[ 16], 00:37:54.112 | 30.00th=[ 19], 40.00th=[ 21], 50.00th=[ 23], 60.00th=[ 28], 00:37:54.112 | 70.00th=[ 33], 80.00th=[ 47], 90.00th=[ 55], 95.00th=[ 108], 00:37:54.112 | 99.00th=[ 178], 99.50th=[ 188], 99.90th=[ 190], 99.95th=[ 190], 00:37:54.112 | 99.99th=[ 190] 00:37:54.112 bw ( KiB/s): min= 9288, max= 9344, per=16.75%, avg=9316.00, stdev=39.60, samples=2 00:37:54.112 iops : min= 2322, max= 2336, avg=2329.00, stdev= 9.90, samples=2 00:37:54.112 lat (msec) : 10=2.04%, 20=37.80%, 50=49.08%, 100=8.26%, 250=2.82% 00:37:54.112 cpu : usr=1.58%, sys=3.35%, ctx=169, majf=0, minf=1 00:37:54.112 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.4%, 32=0.7%, >=64=98.6% 00:37:54.112 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:54.112 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:37:54.112 issued rwts: total=2048,2457,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:54.112 latency : target=0, window=0, percentile=100.00%, depth=128 00:37:54.112 job1: (groupid=0, jobs=1): err= 0: pid=2164852: Wed Nov 6 09:13:44 2024 00:37:54.112 read: IOPS=5034, BW=19.7MiB/s (20.6MB/s)(20.0MiB/1017msec) 00:37:54.112 slat (nsec): min=901, max=20773k, avg=39827.41, stdev=498560.75 00:37:54.112 clat (usec): min=675, max=49320, avg=7314.62, stdev=7719.06 00:37:54.112 lat (usec): min=701, max=49329, avg=7354.45, stdev=7761.36 00:37:54.112 clat percentiles (usec): 00:37:54.112 | 1.00th=[ 1037], 5.00th=[ 1467], 10.00th=[ 2999], 20.00th=[ 3556], 00:37:54.112 | 30.00th=[ 4113], 40.00th=[ 4686], 50.00th=[ 5211], 60.00th=[ 5669], 00:37:54.112 | 70.00th=[ 6259], 80.00th=[ 7439], 90.00th=[14484], 95.00th=[25560], 00:37:54.112 | 99.00th=[47973], 99.50th=[48497], 99.90th=[48497], 99.95th=[48497], 00:37:54.112 | 99.99th=[49546] 00:37:54.112 write: IOPS=6210, BW=24.3MiB/s (25.4MB/s)(24.7MiB/1017msec); 0 zone resets 00:37:54.112 slat (nsec): min=1500, max=22220k, avg=92547.45, stdev=834031.09 00:37:54.112 clat (usec): min=398, max=133316, avg=14518.73, stdev=29858.31 00:37:54.112 lat (usec): min=429, max=133325, avg=14611.28, stdev=30049.45 00:37:54.112 clat percentiles (usec): 00:37:54.112 | 1.00th=[ 1139], 5.00th=[ 1614], 10.00th=[ 2245], 20.00th=[ 2966], 00:37:54.112 | 30.00th=[ 3261], 40.00th=[ 4293], 50.00th=[ 4752], 60.00th=[ 5473], 00:37:54.112 | 70.00th=[ 5669], 80.00th=[ 6980], 90.00th=[ 25297], 95.00th=[110625], 00:37:54.112 | 99.00th=[126354], 99.50th=[128451], 99.90th=[132645], 99.95th=[133694], 00:37:54.112 | 99.99th=[133694] 00:37:54.112 bw ( KiB/s): min= 6888, max=43040, per=44.89%, avg=24964.00, stdev=25563.32, samples=2 00:37:54.112 iops : min= 1722, max=10760, avg=6241.00, stdev=6390.83, samples=2 00:37:54.112 lat (usec) : 500=0.03%, 750=0.17%, 1000=0.45% 00:37:54.112 lat (msec) : 2=6.91%, 4=26.06%, 10=52.51%, 20=2.40%, 50=6.65% 00:37:54.112 lat (msec) : 100=1.06%, 250=3.75% 00:37:54.112 cpu : usr=4.53%, sys=6.00%, ctx=601, majf=0, minf=2 00:37:54.112 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:37:54.112 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:54.112 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:37:54.112 issued rwts: total=5120,6316,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:54.112 latency : target=0, window=0, percentile=100.00%, depth=128 00:37:54.112 job2: (groupid=0, jobs=1): err= 0: pid=2164859: Wed Nov 6 09:13:44 2024 00:37:54.112 read: IOPS=3548, BW=13.9MiB/s (14.5MB/s)(14.0MiB/1010msec) 00:37:54.112 slat (nsec): min=1051, max=17193k, avg=127956.31, stdev=969914.16 00:37:54.112 clat (usec): min=5720, max=46200, avg=17120.33, stdev=6718.29 00:37:54.112 lat (usec): min=5726, max=47697, avg=17248.29, stdev=6777.04 00:37:54.112 clat percentiles (usec): 00:37:54.112 | 1.00th=[ 7504], 5.00th=[ 8848], 10.00th=[ 9896], 20.00th=[11207], 00:37:54.112 | 30.00th=[12518], 40.00th=[13960], 50.00th=[16188], 60.00th=[17957], 00:37:54.112 | 70.00th=[19530], 80.00th=[21365], 90.00th=[27132], 95.00th=[30016], 00:37:54.112 | 99.00th=[38011], 99.50th=[38536], 99.90th=[46400], 99.95th=[46400], 00:37:54.112 | 99.99th=[46400] 00:37:54.112 write: IOPS=3666, BW=14.3MiB/s (15.0MB/s)(14.5MiB/1010msec); 0 zone resets 00:37:54.112 slat (nsec): min=1783, max=21976k, avg=141335.97, stdev=1018741.43 00:37:54.112 clat (usec): min=1202, max=92051, avg=17988.73, stdev=15460.85 00:37:54.112 lat (usec): min=1214, max=92055, avg=18130.06, stdev=15563.53 00:37:54.112 clat percentiles (usec): 00:37:54.112 | 1.00th=[ 6521], 5.00th=[ 7177], 10.00th=[ 7701], 20.00th=[ 8717], 00:37:54.112 | 30.00th=[10683], 40.00th=[11207], 50.00th=[12387], 60.00th=[14353], 00:37:54.112 | 70.00th=[15533], 80.00th=[23462], 90.00th=[36963], 95.00th=[45351], 00:37:54.112 | 99.00th=[88605], 99.50th=[91751], 99.90th=[91751], 99.95th=[91751], 00:37:54.112 | 99.99th=[91751] 00:37:54.112 bw ( KiB/s): min=12296, max=16432, per=25.83%, avg=14364.00, stdev=2924.59, samples=2 00:37:54.112 iops : min= 3074, max= 4108, avg=3591.00, stdev=731.15, samples=2 00:37:54.112 lat (msec) : 2=0.03%, 4=0.01%, 10=18.75%, 20=54.33%, 50=24.59% 00:37:54.112 lat (msec) : 100=2.29% 00:37:54.112 cpu : usr=2.58%, sys=5.15%, ctx=198, majf=0, minf=1 00:37:54.112 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.1% 00:37:54.112 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:54.112 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:37:54.112 issued rwts: total=3584,3703,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:54.112 latency : target=0, window=0, percentile=100.00%, depth=128 00:37:54.112 job3: (groupid=0, jobs=1): err= 0: pid=2164865: Wed Nov 6 09:13:44 2024 00:37:54.112 read: IOPS=1519, BW=6077KiB/s (6223kB/s)(6144KiB/1011msec) 00:37:54.112 slat (nsec): min=1191, max=13964k, avg=253752.26, stdev=1431293.63 00:37:54.112 clat (msec): min=8, max=109, avg=25.49, stdev=18.33 00:37:54.112 lat (msec): min=8, max=109, avg=25.74, stdev=18.52 00:37:54.112 clat percentiles (msec): 00:37:54.112 | 1.00th=[ 11], 5.00th=[ 12], 10.00th=[ 13], 20.00th=[ 16], 00:37:54.112 | 30.00th=[ 17], 40.00th=[ 18], 50.00th=[ 20], 60.00th=[ 22], 00:37:54.112 | 70.00th=[ 27], 80.00th=[ 30], 90.00th=[ 40], 95.00th=[ 73], 00:37:54.112 | 99.00th=[ 102], 99.50th=[ 108], 99.90th=[ 110], 99.95th=[ 110], 00:37:54.112 | 99.99th=[ 110] 00:37:54.112 write: IOPS=1643, BW=6576KiB/s (6733kB/s)(6648KiB/1011msec); 0 zone resets 00:37:54.112 slat (nsec): min=1717, max=22667k, avg=362640.81, stdev=1635087.44 00:37:54.112 clat (msec): min=5, max=133, avg=53.45, stdev=39.97 00:37:54.112 lat (msec): min=8, max=133, avg=53.81, stdev=40.24 00:37:54.112 clat percentiles (msec): 00:37:54.112 | 1.00th=[ 13], 5.00th=[ 13], 10.00th=[ 16], 20.00th=[ 19], 00:37:54.112 | 30.00th=[ 22], 40.00th=[ 33], 50.00th=[ 40], 60.00th=[ 43], 00:37:54.113 | 70.00th=[ 71], 80.00th=[ 108], 90.00th=[ 122], 95.00th=[ 125], 00:37:54.113 | 99.00th=[ 132], 99.50th=[ 133], 99.90th=[ 134], 99.95th=[ 134], 00:37:54.113 | 99.99th=[ 134] 00:37:54.113 bw ( KiB/s): min= 5512, max= 6832, per=11.10%, avg=6172.00, stdev=933.38, samples=2 00:37:54.113 iops : min= 1378, max= 1708, avg=1543.00, stdev=233.35, samples=2 00:37:54.113 lat (msec) : 10=0.31%, 20=38.43%, 50=39.84%, 100=8.26%, 250=13.16% 00:37:54.113 cpu : usr=1.58%, sys=2.18%, ctx=176, majf=0, minf=1 00:37:54.113 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.3%, 16=0.5%, 32=1.0%, >=64=98.0% 00:37:54.113 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:54.113 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:37:54.113 issued rwts: total=1536,1662,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:54.113 latency : target=0, window=0, percentile=100.00%, depth=128 00:37:54.113 00:37:54.113 Run status group 0 (all jobs): 00:37:54.113 READ: bw=47.2MiB/s (49.5MB/s), 6077KiB/s-19.7MiB/s (6223kB/s-20.6MB/s), io=48.0MiB (50.3MB), run=1010-1017msec 00:37:54.113 WRITE: bw=54.3MiB/s (56.9MB/s), 6576KiB/s-24.3MiB/s (6733kB/s-25.4MB/s), io=55.2MiB (57.9MB), run=1010-1017msec 00:37:54.113 00:37:54.113 Disk stats (read/write): 00:37:54.113 nvme0n1: ios=2092/2151, merge=0/0, ticks=26113/27052, in_queue=53165, util=91.58% 00:37:54.113 nvme0n2: ios=3503/4608, merge=0/0, ticks=27702/75781, in_queue=103483, util=87.16% 00:37:54.113 nvme0n3: ios=2801/3072, merge=0/0, ticks=46316/56461, in_queue=102777, util=94.73% 00:37:54.113 nvme0n4: ios=1082/1431, merge=0/0, ticks=23760/78104, in_queue=101864, util=96.58% 00:37:54.113 09:13:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@55 -- # sync 00:37:54.374 09:13:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@59 -- # fio_pid=2165164 00:37:54.374 09:13:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t read -r 10 00:37:54.374 09:13:44 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@61 -- # sleep 3 00:37:54.374 [global] 00:37:54.374 thread=1 00:37:54.374 invalidate=1 00:37:54.374 rw=read 00:37:54.374 time_based=1 00:37:54.374 runtime=10 00:37:54.374 ioengine=libaio 00:37:54.374 direct=1 00:37:54.374 bs=4096 00:37:54.374 iodepth=1 00:37:54.374 norandommap=1 00:37:54.374 numjobs=1 00:37:54.374 00:37:54.374 [job0] 00:37:54.374 filename=/dev/nvme0n1 00:37:54.374 [job1] 00:37:54.374 filename=/dev/nvme0n2 00:37:54.374 [job2] 00:37:54.374 filename=/dev/nvme0n3 00:37:54.374 [job3] 00:37:54.374 filename=/dev/nvme0n4 00:37:54.374 Could not set queue depth (nvme0n1) 00:37:54.374 Could not set queue depth (nvme0n2) 00:37:54.374 Could not set queue depth (nvme0n3) 00:37:54.374 Could not set queue depth (nvme0n4) 00:37:54.635 job0: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:54.635 job1: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:54.635 job2: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:54.635 job3: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:37:54.635 fio-3.35 00:37:54.635 Starting 4 threads 00:37:57.182 09:13:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete concat0 00:37:57.444 fio: io_u error on file /dev/nvme0n4: Operation not supported: read offset=15958016, buflen=4096 00:37:57.444 fio: pid=2165369, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:37:57.444 09:13:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete raid0 00:37:57.704 09:13:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:37:57.705 09:13:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc0 00:37:57.705 fio: io_u error on file /dev/nvme0n3: Operation not supported: read offset=278528, buflen=4096 00:37:57.705 fio: pid=2165365, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:37:57.705 09:13:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:37:57.705 09:13:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc1 00:37:57.705 fio: io_u error on file /dev/nvme0n1: Input/output error: read offset=14606336, buflen=4096 00:37:57.705 fio: pid=2165348, err=5/file:io_u.c:1889, func=io_u error, error=Input/output error 00:37:57.966 09:13:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:37:57.966 09:13:47 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc2 00:37:57.966 fio: io_u error on file /dev/nvme0n2: Operation not supported: read offset=311296, buflen=4096 00:37:57.966 fio: pid=2165357, err=95/file:io_u.c:1889, func=io_u error, error=Operation not supported 00:37:57.966 00:37:57.966 job0: (groupid=0, jobs=1): err= 5 (file:io_u.c:1889, func=io_u error, error=Input/output error): pid=2165348: Wed Nov 6 09:13:48 2024 00:37:57.966 read: IOPS=1198, BW=4791KiB/s (4906kB/s)(13.9MiB/2977msec) 00:37:57.966 slat (usec): min=6, max=8532, avg=29.28, stdev=202.41 00:37:57.966 clat (usec): min=174, max=41216, avg=799.79, stdev=1510.89 00:37:57.966 lat (usec): min=181, max=41241, avg=826.68, stdev=1518.09 00:37:57.966 clat percentiles (usec): 00:37:57.966 | 1.00th=[ 457], 5.00th=[ 562], 10.00th=[ 594], 20.00th=[ 660], 00:37:57.966 | 30.00th=[ 693], 40.00th=[ 717], 50.00th=[ 758], 60.00th=[ 791], 00:37:57.966 | 70.00th=[ 816], 80.00th=[ 840], 90.00th=[ 865], 95.00th=[ 889], 00:37:57.966 | 99.00th=[ 955], 99.50th=[ 988], 99.90th=[41157], 99.95th=[41157], 00:37:57.966 | 99.99th=[41157] 00:37:57.966 bw ( KiB/s): min= 5128, max= 5184, per=53.67%, avg=5166.40, stdev=23.60, samples=5 00:37:57.966 iops : min= 1282, max= 1296, avg=1291.60, stdev= 5.90, samples=5 00:37:57.966 lat (usec) : 250=0.14%, 500=2.10%, 750=46.76%, 1000=50.60% 00:37:57.966 lat (msec) : 2=0.22%, 50=0.14% 00:37:57.966 cpu : usr=1.28%, sys=3.39%, ctx=3570, majf=0, minf=1 00:37:57.966 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:57.966 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.966 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.966 issued rwts: total=3567,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.966 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:57.966 job1: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=2165357: Wed Nov 6 09:13:48 2024 00:37:57.966 read: IOPS=24, BW=96.2KiB/s (98.5kB/s)(304KiB/3161msec) 00:37:57.966 slat (usec): min=18, max=8628, avg=141.90, stdev=980.34 00:37:57.966 clat (usec): min=941, max=42297, avg=41149.95, stdev=4695.78 00:37:57.966 lat (usec): min=975, max=50925, avg=41293.38, stdev=4825.30 00:37:57.966 clat percentiles (usec): 00:37:57.966 | 1.00th=[ 938], 5.00th=[40633], 10.00th=[41157], 20.00th=[41157], 00:37:57.966 | 30.00th=[41681], 40.00th=[41681], 50.00th=[41681], 60.00th=[42206], 00:37:57.966 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:37:57.966 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:37:57.966 | 99.99th=[42206] 00:37:57.966 bw ( KiB/s): min= 89, max= 104, per=1.00%, avg=96.17, stdev= 4.75, samples=6 00:37:57.966 iops : min= 22, max= 26, avg=24.00, stdev= 1.26, samples=6 00:37:57.966 lat (usec) : 1000=1.30% 00:37:57.966 lat (msec) : 50=97.40% 00:37:57.966 cpu : usr=0.09%, sys=0.00%, ctx=80, majf=0, minf=2 00:37:57.966 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:57.966 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.966 complete : 0=1.3%, 4=98.7%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.966 issued rwts: total=77,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.966 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:57.966 job2: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=2165365: Wed Nov 6 09:13:48 2024 00:37:57.966 read: IOPS=24, BW=96.4KiB/s (98.7kB/s)(272KiB/2823msec) 00:37:57.966 slat (nsec): min=25984, max=91705, avg=27797.41, stdev=8015.69 00:37:57.966 clat (usec): min=964, max=42172, avg=41171.74, stdev=4963.58 00:37:57.966 lat (usec): min=1005, max=42199, avg=41199.55, stdev=4962.01 00:37:57.966 clat percentiles (usec): 00:37:57.966 | 1.00th=[ 963], 5.00th=[40633], 10.00th=[41157], 20.00th=[41157], 00:37:57.966 | 30.00th=[41681], 40.00th=[41681], 50.00th=[41681], 60.00th=[42206], 00:37:57.966 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:37:57.966 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:37:57.966 | 99.99th=[42206] 00:37:57.966 bw ( KiB/s): min= 96, max= 96, per=1.00%, avg=96.00, stdev= 0.00, samples=5 00:37:57.966 iops : min= 24, max= 24, avg=24.00, stdev= 0.00, samples=5 00:37:57.966 lat (usec) : 1000=1.45% 00:37:57.966 lat (msec) : 50=97.10% 00:37:57.966 cpu : usr=0.14%, sys=0.00%, ctx=70, majf=0, minf=2 00:37:57.966 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:57.966 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.966 complete : 0=1.4%, 4=98.6%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.966 issued rwts: total=69,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.967 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:57.967 job3: (groupid=0, jobs=1): err=95 (file:io_u.c:1889, func=io_u error, error=Operation not supported): pid=2165369: Wed Nov 6 09:13:48 2024 00:37:57.967 read: IOPS=1498, BW=5994KiB/s (6138kB/s)(15.2MiB/2600msec) 00:37:57.967 slat (nsec): min=6375, max=64605, avg=24751.45, stdev=8014.74 00:37:57.967 clat (usec): min=186, max=982, avg=631.40, stdev=118.47 00:37:57.967 lat (usec): min=194, max=1009, avg=656.15, stdev=120.02 00:37:57.967 clat percentiles (usec): 00:37:57.967 | 1.00th=[ 351], 5.00th=[ 453], 10.00th=[ 478], 20.00th=[ 537], 00:37:57.967 | 30.00th=[ 562], 40.00th=[ 594], 50.00th=[ 627], 60.00th=[ 668], 00:37:57.967 | 70.00th=[ 701], 80.00th=[ 742], 90.00th=[ 791], 95.00th=[ 816], 00:37:57.967 | 99.00th=[ 865], 99.50th=[ 881], 99.90th=[ 938], 99.95th=[ 955], 00:37:57.967 | 99.99th=[ 979] 00:37:57.967 bw ( KiB/s): min= 5664, max= 6928, per=62.95%, avg=6059.20, stdev=527.23, samples=5 00:37:57.967 iops : min= 1416, max= 1732, avg=1514.80, stdev=131.81, samples=5 00:37:57.967 lat (usec) : 250=0.15%, 500=14.04%, 750=66.79%, 1000=18.99% 00:37:57.967 cpu : usr=2.04%, sys=5.89%, ctx=3898, majf=0, minf=2 00:37:57.967 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:37:57.967 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.967 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:37:57.967 issued rwts: total=3897,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:37:57.967 latency : target=0, window=0, percentile=100.00%, depth=1 00:37:57.967 00:37:57.967 Run status group 0 (all jobs): 00:37:57.967 READ: bw=9625KiB/s (9856kB/s), 96.2KiB/s-5994KiB/s (98.5kB/s-6138kB/s), io=29.7MiB (31.2MB), run=2600-3161msec 00:37:57.967 00:37:57.967 Disk stats (read/write): 00:37:57.967 nvme0n1: ios=3562/0, merge=0/0, ticks=2587/0, in_queue=2587, util=94.39% 00:37:57.967 nvme0n2: ios=74/0, merge=0/0, ticks=3046/0, in_queue=3046, util=95.66% 00:37:57.967 nvme0n3: ios=62/0, merge=0/0, ticks=2551/0, in_queue=2551, util=96.03% 00:37:57.967 nvme0n4: ios=3897/0, merge=0/0, ticks=2099/0, in_queue=2099, util=96.42% 00:37:58.228 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:37:58.228 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc3 00:37:58.228 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:37:58.228 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc4 00:37:58.526 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:37:58.526 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc5 00:37:58.831 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:37:58.831 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc6 00:37:58.831 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@69 -- # fio_status=0 00:37:58.831 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@70 -- # wait 2165164 00:37:58.831 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@70 -- # fio_status=4 00:37:58.831 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@72 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:37:59.129 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:37:59.129 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@73 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:37:59.129 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1219 -- # local i=0 00:37:59.129 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:37:59.129 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:37:59.129 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:37:59.129 09:13:48 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1231 -- # return 0 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@75 -- # '[' 4 -eq 0 ']' 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@80 -- # echo 'nvmf hotplug test: fio failed as expected' 00:37:59.129 nvmf hotplug test: fio failed as expected 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@85 -- # rm -f ./local-job0-0-verify.state 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@86 -- # rm -f ./local-job1-1-verify.state 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@87 -- # rm -f ./local-job2-2-verify.state 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@89 -- # trap - SIGINT SIGTERM EXIT 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- target/fio.sh@91 -- # nvmftestfini 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@514 -- # nvmfcleanup 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@121 -- # sync 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@124 -- # set +e 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@125 -- # for i in {1..20} 00:37:59.129 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:37:59.129 rmmod nvme_tcp 00:37:59.403 rmmod nvme_fabrics 00:37:59.403 rmmod nvme_keyring 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@128 -- # set -e 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@129 -- # return 0 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@515 -- # '[' -n 2161990 ']' 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@516 -- # killprocess 2161990 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@950 -- # '[' -z 2161990 ']' 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@954 -- # kill -0 2161990 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@955 -- # uname 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2161990 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2161990' 00:37:59.403 killing process with pid 2161990 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@969 -- # kill 2161990 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@974 -- # wait 2161990 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@297 -- # iptr 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@789 -- # iptables-save 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@789 -- # iptables-restore 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@302 -- # remove_spdk_ns 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:37:59.403 09:13:49 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:01.946 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:38:01.946 00:38:01.946 real 0m27.689s 00:38:01.946 user 2m20.820s 00:38:01.946 sys 0m12.132s 00:38:01.946 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@1126 -- # xtrace_disable 00:38:01.946 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:38:01.946 ************************************ 00:38:01.946 END TEST nvmf_fio_target 00:38:01.946 ************************************ 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@35 -- # run_test nvmf_bdevio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --interrupt-mode 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1107 -- # xtrace_disable 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:38:01.947 ************************************ 00:38:01.947 START TEST nvmf_bdevio 00:38:01.947 ************************************ 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --interrupt-mode 00:38:01.947 * Looking for test storage... 00:38:01.947 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1689 -- # lcov --version 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@333 -- # local ver1 ver1_l 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@334 -- # local ver2 ver2_l 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@336 -- # IFS=.-: 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@336 -- # read -ra ver1 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@337 -- # IFS=.-: 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@337 -- # read -ra ver2 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@338 -- # local 'op=<' 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@340 -- # ver1_l=2 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@341 -- # ver2_l=1 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@344 -- # case "$op" in 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@345 -- # : 1 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@364 -- # (( v = 0 )) 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@365 -- # decimal 1 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@353 -- # local d=1 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@355 -- # echo 1 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@365 -- # ver1[v]=1 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@366 -- # decimal 2 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@353 -- # local d=2 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@355 -- # echo 2 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@366 -- # ver2[v]=2 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@368 -- # return 0 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:38:01.947 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:01.947 --rc genhtml_branch_coverage=1 00:38:01.947 --rc genhtml_function_coverage=1 00:38:01.947 --rc genhtml_legend=1 00:38:01.947 --rc geninfo_all_blocks=1 00:38:01.947 --rc geninfo_unexecuted_blocks=1 00:38:01.947 00:38:01.947 ' 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:38:01.947 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:01.947 --rc genhtml_branch_coverage=1 00:38:01.947 --rc genhtml_function_coverage=1 00:38:01.947 --rc genhtml_legend=1 00:38:01.947 --rc geninfo_all_blocks=1 00:38:01.947 --rc geninfo_unexecuted_blocks=1 00:38:01.947 00:38:01.947 ' 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:38:01.947 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:01.947 --rc genhtml_branch_coverage=1 00:38:01.947 --rc genhtml_function_coverage=1 00:38:01.947 --rc genhtml_legend=1 00:38:01.947 --rc geninfo_all_blocks=1 00:38:01.947 --rc geninfo_unexecuted_blocks=1 00:38:01.947 00:38:01.947 ' 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:38:01.947 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:01.947 --rc genhtml_branch_coverage=1 00:38:01.947 --rc genhtml_function_coverage=1 00:38:01.947 --rc genhtml_legend=1 00:38:01.947 --rc geninfo_all_blocks=1 00:38:01.947 --rc geninfo_unexecuted_blocks=1 00:38:01.947 00:38:01.947 ' 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@7 -- # uname -s 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@15 -- # shopt -s extglob 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:01.947 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@5 -- # export PATH 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@51 -- # : 0 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@55 -- # have_pci_nics=0 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@14 -- # nvmftestinit 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@474 -- # prepare_net_devs 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@436 -- # local -g is_hw=no 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@438 -- # remove_spdk_ns 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@309 -- # xtrace_disable 00:38:01.948 09:13:51 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@315 -- # pci_devs=() 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@315 -- # local -a pci_devs 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@316 -- # pci_net_devs=() 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@317 -- # pci_drivers=() 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@317 -- # local -A pci_drivers 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@319 -- # net_devs=() 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@319 -- # local -ga net_devs 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@320 -- # e810=() 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@320 -- # local -ga e810 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@321 -- # x722=() 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@321 -- # local -ga x722 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@322 -- # mlx=() 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@322 -- # local -ga mlx 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:38:10.091 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:38:10.091 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ up == up ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:38:10.091 Found net devices under 0000:4b:00.0: cvl_0_0 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ up == up ]] 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:38:10.091 Found net devices under 0000:4b:00.1: cvl_0_1 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:38:10.091 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@440 -- # is_hw=yes 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:38:10.092 09:13:58 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:38:10.092 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:38:10.092 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.556 ms 00:38:10.092 00:38:10.092 --- 10.0.0.2 ping statistics --- 00:38:10.092 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:10.092 rtt min/avg/max/mdev = 0.556/0.556/0.556/0.000 ms 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:38:10.092 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:38:10.092 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.275 ms 00:38:10.092 00:38:10.092 --- 10.0.0.1 ping statistics --- 00:38:10.092 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:10.092 rtt min/avg/max/mdev = 0.275/0.275/0.275/0.000 ms 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@448 -- # return 0 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@724 -- # xtrace_disable 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@507 -- # nvmfpid=2170392 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@508 -- # waitforlisten 2170392 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x78 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@831 -- # '[' -z 2170392 ']' 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@836 -- # local max_retries=100 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:38:10.092 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@840 -- # xtrace_disable 00:38:10.092 09:13:59 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:10.092 [2024-11-06 09:13:59.222985] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:38:10.092 [2024-11-06 09:13:59.224166] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:38:10.092 [2024-11-06 09:13:59.224221] [ DPDK EAL parameters: nvmf -c 0x78 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:38:10.092 [2024-11-06 09:13:59.323945] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:38:10.092 [2024-11-06 09:13:59.376787] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:38:10.092 [2024-11-06 09:13:59.376840] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:38:10.092 [2024-11-06 09:13:59.376850] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:38:10.092 [2024-11-06 09:13:59.376857] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:38:10.092 [2024-11-06 09:13:59.376865] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:38:10.092 [2024-11-06 09:13:59.379173] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 4 00:38:10.092 [2024-11-06 09:13:59.379339] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 5 00:38:10.092 [2024-11-06 09:13:59.379514] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 6 00:38:10.092 [2024-11-06 09:13:59.379514] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:38:10.092 [2024-11-06 09:13:59.456053] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:38:10.092 [2024-11-06 09:13:59.457252] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:38:10.092 [2024-11-06 09:13:59.457258] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:38:10.092 [2024-11-06 09:13:59.457777] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:38:10.092 [2024-11-06 09:13:59.457831] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@864 -- # return 0 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@730 -- # xtrace_disable 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:10.092 [2024-11-06 09:14:00.084549] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:10.092 Malloc0 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:10.092 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:10.093 [2024-11-06 09:14:00.188948] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@558 -- # config=() 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@558 -- # local subsystem config 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:38:10.093 { 00:38:10.093 "params": { 00:38:10.093 "name": "Nvme$subsystem", 00:38:10.093 "trtype": "$TEST_TRANSPORT", 00:38:10.093 "traddr": "$NVMF_FIRST_TARGET_IP", 00:38:10.093 "adrfam": "ipv4", 00:38:10.093 "trsvcid": "$NVMF_PORT", 00:38:10.093 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:38:10.093 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:38:10.093 "hdgst": ${hdgst:-false}, 00:38:10.093 "ddgst": ${ddgst:-false} 00:38:10.093 }, 00:38:10.093 "method": "bdev_nvme_attach_controller" 00:38:10.093 } 00:38:10.093 EOF 00:38:10.093 )") 00:38:10.093 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@580 -- # cat 00:38:10.353 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@582 -- # jq . 00:38:10.353 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@583 -- # IFS=, 00:38:10.353 09:14:00 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:38:10.353 "params": { 00:38:10.353 "name": "Nvme1", 00:38:10.353 "trtype": "tcp", 00:38:10.353 "traddr": "10.0.0.2", 00:38:10.353 "adrfam": "ipv4", 00:38:10.353 "trsvcid": "4420", 00:38:10.353 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:38:10.353 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:38:10.353 "hdgst": false, 00:38:10.353 "ddgst": false 00:38:10.353 }, 00:38:10.353 "method": "bdev_nvme_attach_controller" 00:38:10.353 }' 00:38:10.353 [2024-11-06 09:14:00.247500] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:38:10.354 [2024-11-06 09:14:00.247573] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2170737 ] 00:38:10.354 [2024-11-06 09:14:00.324204] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 3 00:38:10.354 [2024-11-06 09:14:00.369184] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:10.354 [2024-11-06 09:14:00.369304] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:38:10.354 [2024-11-06 09:14:00.369307] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:38:10.614 I/O targets: 00:38:10.614 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:38:10.614 00:38:10.614 00:38:10.614 CUnit - A unit testing framework for C - Version 2.1-3 00:38:10.614 http://cunit.sourceforge.net/ 00:38:10.614 00:38:10.614 00:38:10.614 Suite: bdevio tests on: Nvme1n1 00:38:10.874 Test: blockdev write read block ...passed 00:38:10.874 Test: blockdev write zeroes read block ...passed 00:38:10.874 Test: blockdev write zeroes read no split ...passed 00:38:10.874 Test: blockdev write zeroes read split ...passed 00:38:10.874 Test: blockdev write zeroes read split partial ...passed 00:38:10.874 Test: blockdev reset ...[2024-11-06 09:14:00.869086] nvme_ctrlr.c:1701:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 1] resetting controller 00:38:10.874 [2024-11-06 09:14:00.869157] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xb6a970 (9): Bad file descriptor 00:38:10.874 [2024-11-06 09:14:00.916645] bdev_nvme.c:2236:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: [nqn.2016-06.io.spdk:cnode1, 2] Resetting controller successful. 00:38:10.874 passed 00:38:10.874 Test: blockdev write read 8 blocks ...passed 00:38:10.874 Test: blockdev write read size > 128k ...passed 00:38:10.874 Test: blockdev write read invalid size ...passed 00:38:11.133 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:38:11.133 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:38:11.133 Test: blockdev write read max offset ...passed 00:38:11.133 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:38:11.133 Test: blockdev writev readv 8 blocks ...passed 00:38:11.133 Test: blockdev writev readv 30 x 1block ...passed 00:38:11.133 Test: blockdev writev readv block ...passed 00:38:11.133 Test: blockdev writev readv size > 128k ...passed 00:38:11.133 Test: blockdev writev readv size > 128k in two iovs ...passed 00:38:11.133 Test: blockdev comparev and writev ...[2024-11-06 09:14:01.179636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:38:11.133 [2024-11-06 09:14:01.179662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:38:11.133 [2024-11-06 09:14:01.179673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:38:11.133 [2024-11-06 09:14:01.179678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:38:11.133 [2024-11-06 09:14:01.180187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:38:11.133 [2024-11-06 09:14:01.180198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:38:11.133 [2024-11-06 09:14:01.180210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:38:11.133 [2024-11-06 09:14:01.180215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:38:11.133 [2024-11-06 09:14:01.180757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:38:11.133 [2024-11-06 09:14:01.180766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:38:11.133 [2024-11-06 09:14:01.180776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:38:11.133 [2024-11-06 09:14:01.180782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:38:11.133 [2024-11-06 09:14:01.181311] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:38:11.133 [2024-11-06 09:14:01.181320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:38:11.133 [2024-11-06 09:14:01.181331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:38:11.133 [2024-11-06 09:14:01.181336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:38:11.133 passed 00:38:11.393 Test: blockdev nvme passthru rw ...passed 00:38:11.393 Test: blockdev nvme passthru vendor specific ...[2024-11-06 09:14:01.265589] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:38:11.393 [2024-11-06 09:14:01.265600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:38:11.393 [2024-11-06 09:14:01.265961] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:38:11.393 [2024-11-06 09:14:01.265970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:38:11.393 [2024-11-06 09:14:01.266326] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:38:11.393 [2024-11-06 09:14:01.266334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:38:11.393 [2024-11-06 09:14:01.266660] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:38:11.393 [2024-11-06 09:14:01.266669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:38:11.393 passed 00:38:11.393 Test: blockdev nvme admin passthru ...passed 00:38:11.393 Test: blockdev copy ...passed 00:38:11.393 00:38:11.393 Run Summary: Type Total Ran Passed Failed Inactive 00:38:11.393 suites 1 1 n/a 0 0 00:38:11.393 tests 23 23 23 0 0 00:38:11.393 asserts 152 152 152 0 n/a 00:38:11.393 00:38:11.393 Elapsed time = 1.225 seconds 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- target/bdevio.sh@30 -- # nvmftestfini 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@514 -- # nvmfcleanup 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@121 -- # sync 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@124 -- # set +e 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@125 -- # for i in {1..20} 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:38:11.393 rmmod nvme_tcp 00:38:11.393 rmmod nvme_fabrics 00:38:11.393 rmmod nvme_keyring 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@128 -- # set -e 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@129 -- # return 0 00:38:11.393 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@515 -- # '[' -n 2170392 ']' 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@516 -- # killprocess 2170392 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@950 -- # '[' -z 2170392 ']' 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@954 -- # kill -0 2170392 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@955 -- # uname 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2170392 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@956 -- # process_name=reactor_3 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@960 -- # '[' reactor_3 = sudo ']' 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2170392' 00:38:11.654 killing process with pid 2170392 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@969 -- # kill 2170392 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@974 -- # wait 2170392 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@297 -- # iptr 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@789 -- # iptables-save 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@789 -- # iptables-restore 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@302 -- # remove_spdk_ns 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 15> /dev/null' 00:38:11.654 09:14:01 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:14.196 09:14:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:38:14.196 00:38:14.196 real 0m12.190s 00:38:14.196 user 0m10.367s 00:38:14.196 sys 0m6.414s 00:38:14.196 09:14:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@1126 -- # xtrace_disable 00:38:14.196 09:14:03 nvmf_tcp.nvmf_target_core_interrupt_mode.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:38:14.196 ************************************ 00:38:14.196 END TEST nvmf_bdevio 00:38:14.196 ************************************ 00:38:14.196 09:14:03 nvmf_tcp.nvmf_target_core_interrupt_mode -- nvmf/nvmf_target_core.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:38:14.196 00:38:14.196 real 4m56.163s 00:38:14.196 user 10m21.135s 00:38:14.196 sys 2m3.515s 00:38:14.196 09:14:03 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@1126 -- # xtrace_disable 00:38:14.196 09:14:03 nvmf_tcp.nvmf_target_core_interrupt_mode -- common/autotest_common.sh@10 -- # set +x 00:38:14.196 ************************************ 00:38:14.196 END TEST nvmf_target_core_interrupt_mode 00:38:14.196 ************************************ 00:38:14.196 09:14:03 nvmf_tcp -- nvmf/nvmf.sh@21 -- # run_test nvmf_interrupt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/interrupt.sh --transport=tcp --interrupt-mode 00:38:14.196 09:14:03 nvmf_tcp -- common/autotest_common.sh@1101 -- # '[' 4 -le 1 ']' 00:38:14.196 09:14:03 nvmf_tcp -- common/autotest_common.sh@1107 -- # xtrace_disable 00:38:14.196 09:14:03 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:14.196 ************************************ 00:38:14.196 START TEST nvmf_interrupt 00:38:14.196 ************************************ 00:38:14.196 09:14:03 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/interrupt.sh --transport=tcp --interrupt-mode 00:38:14.196 * Looking for test storage... 00:38:14.196 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1689 -- # lcov --version 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@333 -- # local ver1 ver1_l 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@334 -- # local ver2 ver2_l 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@336 -- # IFS=.-: 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@336 -- # read -ra ver1 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@337 -- # IFS=.-: 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@337 -- # read -ra ver2 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@338 -- # local 'op=<' 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@340 -- # ver1_l=2 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@341 -- # ver2_l=1 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@344 -- # case "$op" in 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@345 -- # : 1 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@364 -- # (( v = 0 )) 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@365 -- # decimal 1 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@353 -- # local d=1 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@355 -- # echo 1 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@365 -- # ver1[v]=1 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@366 -- # decimal 2 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@353 -- # local d=2 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@355 -- # echo 2 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@366 -- # ver2[v]=2 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@368 -- # return 0 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:38:14.196 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:14.196 --rc genhtml_branch_coverage=1 00:38:14.196 --rc genhtml_function_coverage=1 00:38:14.196 --rc genhtml_legend=1 00:38:14.196 --rc geninfo_all_blocks=1 00:38:14.196 --rc geninfo_unexecuted_blocks=1 00:38:14.196 00:38:14.196 ' 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:38:14.196 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:14.196 --rc genhtml_branch_coverage=1 00:38:14.196 --rc genhtml_function_coverage=1 00:38:14.196 --rc genhtml_legend=1 00:38:14.196 --rc geninfo_all_blocks=1 00:38:14.196 --rc geninfo_unexecuted_blocks=1 00:38:14.196 00:38:14.196 ' 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:38:14.196 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:14.196 --rc genhtml_branch_coverage=1 00:38:14.196 --rc genhtml_function_coverage=1 00:38:14.196 --rc genhtml_legend=1 00:38:14.196 --rc geninfo_all_blocks=1 00:38:14.196 --rc geninfo_unexecuted_blocks=1 00:38:14.196 00:38:14.196 ' 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:38:14.196 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:14.196 --rc genhtml_branch_coverage=1 00:38:14.196 --rc genhtml_function_coverage=1 00:38:14.196 --rc genhtml_legend=1 00:38:14.196 --rc geninfo_all_blocks=1 00:38:14.196 --rc geninfo_unexecuted_blocks=1 00:38:14.196 00:38:14.196 ' 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@7 -- # uname -s 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@15 -- # shopt -s extglob 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:14.196 09:14:04 nvmf_tcp.nvmf_interrupt -- paths/export.sh@5 -- # export PATH 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@51 -- # : 0 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@33 -- # '[' 1 -eq 1 ']' 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@34 -- # NVMF_APP+=(--interrupt-mode) 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@55 -- # have_pci_nics=0 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/interrupt/common.sh 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@12 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@14 -- # nvmftestinit 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@474 -- # prepare_net_devs 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@436 -- # local -g is_hw=no 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@438 -- # remove_spdk_ns 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@309 -- # xtrace_disable 00:38:14.197 09:14:04 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@315 -- # pci_devs=() 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@315 -- # local -a pci_devs 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@316 -- # pci_net_devs=() 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@317 -- # pci_drivers=() 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@317 -- # local -A pci_drivers 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@319 -- # net_devs=() 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@319 -- # local -ga net_devs 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@320 -- # e810=() 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@320 -- # local -ga e810 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@321 -- # x722=() 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@321 -- # local -ga x722 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@322 -- # mlx=() 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@322 -- # local -ga mlx 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:38:22.335 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:38:22.335 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:22.335 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@416 -- # [[ up == up ]] 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:38:22.336 Found net devices under 0000:4b:00.0: cvl_0_0 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@416 -- # [[ up == up ]] 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:38:22.336 Found net devices under 0000:4b:00.1: cvl_0_1 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@440 -- # is_hw=yes 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:38:22.336 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:38:22.336 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.687 ms 00:38:22.336 00:38:22.336 --- 10.0.0.2 ping statistics --- 00:38:22.336 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:22.336 rtt min/avg/max/mdev = 0.687/0.687/0.687/0.000 ms 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:38:22.336 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:38:22.336 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.308 ms 00:38:22.336 00:38:22.336 --- 10.0.0.1 ping statistics --- 00:38:22.336 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:38:22.336 rtt min/avg/max/mdev = 0.308/0.308/0.308/0.000 ms 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@448 -- # return 0 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@15 -- # nvmfappstart -m 0x3 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@724 -- # xtrace_disable 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@507 -- # nvmfpid=2175091 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@508 -- # waitforlisten 2175091 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --interrupt-mode -m 0x3 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@831 -- # '[' -z 2175091 ']' 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@836 -- # local max_retries=100 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:38:22.336 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@840 -- # xtrace_disable 00:38:22.336 09:14:11 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:38:22.336 [2024-11-06 09:14:11.605596] thread.c:2964:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:38:22.336 [2024-11-06 09:14:11.606759] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:38:22.336 [2024-11-06 09:14:11.606811] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:38:22.336 [2024-11-06 09:14:11.688683] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:38:22.336 [2024-11-06 09:14:11.728902] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:38:22.336 [2024-11-06 09:14:11.728938] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:38:22.336 [2024-11-06 09:14:11.728946] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:38:22.336 [2024-11-06 09:14:11.728953] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:38:22.336 [2024-11-06 09:14:11.728959] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:38:22.336 [2024-11-06 09:14:11.730545] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:22.336 [2024-11-06 09:14:11.730548] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:38:22.336 [2024-11-06 09:14:11.787468] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:38:22.336 [2024-11-06 09:14:11.787937] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:38:22.336 [2024-11-06 09:14:11.788280] thread.c:2115:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:38:22.336 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:38:22.336 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@864 -- # return 0 00:38:22.336 09:14:12 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:38:22.336 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@730 -- # xtrace_disable 00:38:22.336 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:38:22.597 09:14:12 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:38:22.597 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@16 -- # setup_bdev_aio 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@77 -- # uname -s 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@77 -- # [[ Linux != \F\r\e\e\B\S\D ]] 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@78 -- # dd if=/dev/zero of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aiofile bs=2048 count=5000 00:38:22.598 5000+0 records in 00:38:22.598 5000+0 records out 00:38:22.598 10240000 bytes (10 MB, 9.8 MiB) copied, 0.0172015 s, 595 MB/s 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@79 -- # rpc_cmd bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aiofile AIO0 2048 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:38:22.598 AIO0 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -q 256 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:38:22.598 [2024-11-06 09:14:12.531158] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 AIO0 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@561 -- # xtrace_disable 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:38:22.598 [2024-11-06 09:14:12.571411] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@24 -- # for i in {0..1} 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@25 -- # reactor_is_idle 2175091 0 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 2175091 0 idle 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=2175091 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 2175091 -w 256 00:38:22.598 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='2175091 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.24 reactor_0' 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 2175091 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.24 reactor_0 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@24 -- # for i in {0..1} 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@25 -- # reactor_is_idle 2175091 1 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 2175091 1 idle 00:38:22.858 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=2175091 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 2175091 -w 256 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='2175095 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.00 reactor_1' 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 2175095 root 20 0 128.2g 43776 32256 S 0.0 0.0 0:00.00 reactor_1 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@28 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@35 -- # perf_pid=2175453 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@38 -- # for i in {0..1} 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # BUSY_THRESHOLD=30 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 256 -o 4096 -w randrw -M 30 -t 10 -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # reactor_is_busy 2175091 0 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@49 -- # reactor_is_busy_or_idle 2175091 0 busy 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=2175091 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=30 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ busy != \b\u\s\y ]] 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 2175091 -w 256 00:38:22.859 09:14:12 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='2175091 root 20 0 128.2g 44928 32256 R 40.0 0.0 0:00.31 reactor_0' 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 2175091 root 20 0 128.2g 44928 32256 R 40.0 0.0 0:00.31 reactor_0 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=40.0 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=40 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ busy = \b\u\s\y ]] 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # (( cpu_rate < busy_threshold )) 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ busy = \i\d\l\e ]] 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@38 -- # for i in {0..1} 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # BUSY_THRESHOLD=30 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@39 -- # reactor_is_busy 2175091 1 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@49 -- # reactor_is_busy_or_idle 2175091 1 busy 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=2175091 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=busy 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=30 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ busy != \b\u\s\y ]] 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 2175091 -w 256 00:38:23.119 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='2175095 root 20 0 128.2g 44928 32256 R 99.9 0.0 0:00.22 reactor_1' 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 2175095 root 20 0 128.2g 44928 32256 R 99.9 0.0 0:00.22 reactor_1 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=99.9 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=99 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ busy = \b\u\s\y ]] 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # (( cpu_rate < busy_threshold )) 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ busy = \i\d\l\e ]] 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:38:23.380 09:14:13 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@42 -- # wait 2175453 00:38:33.375 Initializing NVMe Controllers 00:38:33.375 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:38:33.375 Controller IO queue size 256, less than required. 00:38:33.375 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:38:33.375 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:38:33.375 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:38:33.375 Initialization complete. Launching workers. 00:38:33.376 ======================================================== 00:38:33.376 Latency(us) 00:38:33.376 Device Information : IOPS MiB/s Average min max 00:38:33.376 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 16493.59 64.43 15530.46 2462.20 18033.71 00:38:33.376 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 19602.09 76.57 13061.92 7485.41 29290.39 00:38:33.376 ======================================================== 00:38:33.376 Total : 36095.68 141.00 14189.89 2462.20 29290.39 00:38:33.376 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@45 -- # for i in {0..1} 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@46 -- # reactor_is_idle 2175091 0 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 2175091 0 idle 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=2175091 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 2175091 -w 256 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='2175091 root 20 0 128.2g 44928 32256 S 0.0 0.0 0:20.25 reactor_0' 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 2175091 root 20 0 128.2g 44928 32256 S 0.0 0.0 0:20.25 reactor_0 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@45 -- # for i in {0..1} 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@46 -- # reactor_is_idle 2175091 1 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 2175091 1 idle 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=2175091 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 2175091 -w 256 00:38:33.376 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='2175095 root 20 0 128.2g 44928 32256 S 0.0 0.0 0:10.00 reactor_1' 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 2175095 root 20 0 128.2g 44928 32256 S 0.0 0.0 0:10.00 reactor_1 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:38:33.637 09:14:23 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@50 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:38:34.207 09:14:24 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@51 -- # waitforserial SPDKISFASTANDAWESOME 00:38:34.207 09:14:24 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1198 -- # local i=0 00:38:34.207 09:14:24 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:38:34.207 09:14:24 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:38:34.207 09:14:24 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1205 -- # sleep 2 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1208 -- # return 0 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@52 -- # for i in {0..1} 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@53 -- # reactor_is_idle 2175091 0 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 2175091 0 idle 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=2175091 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=0 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 2175091 -w 256 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_0 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='2175091 root 20 0 128.2g 79488 32256 R 0.0 0.1 0:20.50 reactor_0' 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 2175091 root 20 0 128.2g 79488 32256 R 0.0 0.1 0:20.50 reactor_0 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:38:36.119 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@52 -- # for i in {0..1} 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@53 -- # reactor_is_idle 2175091 1 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@53 -- # reactor_is_busy_or_idle 2175091 1 idle 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@10 -- # local pid=2175091 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@11 -- # local idx=1 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@12 -- # local state=idle 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@13 -- # local busy_threshold=65 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@14 -- # local idle_threshold=30 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \b\u\s\y ]] 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@16 -- # [[ idle != \i\d\l\e ]] 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@20 -- # hash top 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j = 10 )) 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@25 -- # (( j != 0 )) 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # grep reactor_1 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top -bHn 1 -p 2175091 -w 256 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@26 -- # top_reactor='2175095 root 20 0 128.2g 79488 32256 S 0.0 0.1 0:10.15 reactor_1' 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # echo 2175095 root 20 0 128.2g 79488 32256 S 0.0 0.1 0:10.15 reactor_1 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # sed -e 's/^\s*//g' 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # awk '{print $9}' 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@27 -- # cpu_rate=0.0 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@28 -- # cpu_rate=0 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@30 -- # [[ idle = \b\u\s\y ]] 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # [[ idle = \i\d\l\e ]] 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@32 -- # (( cpu_rate > idle_threshold )) 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- interrupt/common.sh@35 -- # return 0 00:38:36.380 09:14:26 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@55 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:38:36.640 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@56 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1219 -- # local i=0 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1231 -- # return 0 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@58 -- # trap - SIGINT SIGTERM EXIT 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- target/interrupt.sh@59 -- # nvmftestfini 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@514 -- # nvmfcleanup 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@121 -- # sync 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@124 -- # set +e 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@125 -- # for i in {1..20} 00:38:36.640 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:38:36.640 rmmod nvme_tcp 00:38:36.640 rmmod nvme_fabrics 00:38:36.640 rmmod nvme_keyring 00:38:36.900 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:38:36.900 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@128 -- # set -e 00:38:36.900 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@129 -- # return 0 00:38:36.900 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@515 -- # '[' -n 2175091 ']' 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@516 -- # killprocess 2175091 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@950 -- # '[' -z 2175091 ']' 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@954 -- # kill -0 2175091 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@955 -- # uname 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2175091 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2175091' 00:38:36.901 killing process with pid 2175091 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@969 -- # kill 2175091 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@974 -- # wait 2175091 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@297 -- # iptr 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@789 -- # iptables-save 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@789 -- # iptables-restore 00:38:36.901 09:14:26 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:38:36.901 09:14:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@302 -- # remove_spdk_ns 00:38:36.901 09:14:27 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:36.901 09:14:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:38:36.901 09:14:27 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:39.445 09:14:29 nvmf_tcp.nvmf_interrupt -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:38:39.445 00:38:39.445 real 0m25.129s 00:38:39.445 user 0m40.291s 00:38:39.445 sys 0m9.385s 00:38:39.445 09:14:29 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@1126 -- # xtrace_disable 00:38:39.445 09:14:29 nvmf_tcp.nvmf_interrupt -- common/autotest_common.sh@10 -- # set +x 00:38:39.445 ************************************ 00:38:39.445 END TEST nvmf_interrupt 00:38:39.445 ************************************ 00:38:39.445 00:38:39.445 real 29m40.330s 00:38:39.445 user 61m38.734s 00:38:39.445 sys 9m54.657s 00:38:39.445 09:14:29 nvmf_tcp -- common/autotest_common.sh@1126 -- # xtrace_disable 00:38:39.445 09:14:29 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:39.445 ************************************ 00:38:39.445 END TEST nvmf_tcp 00:38:39.445 ************************************ 00:38:39.445 09:14:29 -- spdk/autotest.sh@281 -- # [[ 0 -eq 0 ]] 00:38:39.445 09:14:29 -- spdk/autotest.sh@282 -- # run_test spdkcli_nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:38:39.445 09:14:29 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:38:39.445 09:14:29 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:38:39.445 09:14:29 -- common/autotest_common.sh@10 -- # set +x 00:38:39.445 ************************************ 00:38:39.445 START TEST spdkcli_nvmf_tcp 00:38:39.445 ************************************ 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:38:39.445 * Looking for test storage... 00:38:39.445 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@1689 -- # lcov --version 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@333 -- # local ver1 ver1_l 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@334 -- # local ver2 ver2_l 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@336 -- # IFS=.-: 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@336 -- # read -ra ver1 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@337 -- # IFS=.-: 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@337 -- # read -ra ver2 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@338 -- # local 'op=<' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@340 -- # ver1_l=2 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@341 -- # ver2_l=1 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@344 -- # case "$op" in 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@345 -- # : 1 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@364 -- # (( v = 0 )) 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@365 -- # decimal 1 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@353 -- # local d=1 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@355 -- # echo 1 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@365 -- # ver1[v]=1 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@366 -- # decimal 2 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@353 -- # local d=2 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@355 -- # echo 2 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@366 -- # ver2[v]=2 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@368 -- # return 0 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:38:39.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:39.445 --rc genhtml_branch_coverage=1 00:38:39.445 --rc genhtml_function_coverage=1 00:38:39.445 --rc genhtml_legend=1 00:38:39.445 --rc geninfo_all_blocks=1 00:38:39.445 --rc geninfo_unexecuted_blocks=1 00:38:39.445 00:38:39.445 ' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:38:39.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:39.445 --rc genhtml_branch_coverage=1 00:38:39.445 --rc genhtml_function_coverage=1 00:38:39.445 --rc genhtml_legend=1 00:38:39.445 --rc geninfo_all_blocks=1 00:38:39.445 --rc geninfo_unexecuted_blocks=1 00:38:39.445 00:38:39.445 ' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:38:39.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:39.445 --rc genhtml_branch_coverage=1 00:38:39.445 --rc genhtml_function_coverage=1 00:38:39.445 --rc genhtml_legend=1 00:38:39.445 --rc geninfo_all_blocks=1 00:38:39.445 --rc geninfo_unexecuted_blocks=1 00:38:39.445 00:38:39.445 ' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:38:39.445 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:39.445 --rc genhtml_branch_coverage=1 00:38:39.445 --rc genhtml_function_coverage=1 00:38:39.445 --rc genhtml_legend=1 00:38:39.445 --rc geninfo_all_blocks=1 00:38:39.445 --rc geninfo_unexecuted_blocks=1 00:38:39.445 00:38:39.445 ' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@7 -- # uname -s 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@15 -- # shopt -s extglob 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- paths/export.sh@5 -- # export PATH 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@51 -- # : 0 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:38:39.445 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- nvmf/common.sh@55 -- # have_pci_nics=0 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@12 -- # MATCH_FILE=spdkcli_nvmf.test 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@13 -- # SPDKCLI_BRANCH=/nvmf 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@15 -- # trap cleanup EXIT 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@17 -- # timing_enter run_nvmf_tgt 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@18 -- # run_nvmf_tgt 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/common.sh@33 -- # nvmf_tgt_pid=2178631 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/common.sh@34 -- # waitforlisten 2178631 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@831 -- # '[' -z 2178631 ']' 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- spdkcli/common.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x3 -p 0 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@836 -- # local max_retries=100 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:38:39.445 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@840 -- # xtrace_disable 00:38:39.445 09:14:29 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:39.445 [2024-11-06 09:14:29.492714] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:38:39.445 [2024-11-06 09:14:29.492802] [ DPDK EAL parameters: nvmf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2178631 ] 00:38:39.705 [2024-11-06 09:14:29.570165] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 2 00:38:39.705 [2024-11-06 09:14:29.613172] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:38:39.705 [2024-11-06 09:14:29.613175] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:38:40.275 09:14:30 spdkcli_nvmf_tcp -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:38:40.275 09:14:30 spdkcli_nvmf_tcp -- common/autotest_common.sh@864 -- # return 0 00:38:40.276 09:14:30 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@19 -- # timing_exit run_nvmf_tgt 00:38:40.276 09:14:30 spdkcli_nvmf_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:38:40.276 09:14:30 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:40.276 09:14:30 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@21 -- # NVMF_TARGET_IP=127.0.0.1 00:38:40.276 09:14:30 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@22 -- # [[ tcp == \r\d\m\a ]] 00:38:40.276 09:14:30 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@27 -- # timing_enter spdkcli_create_nvmf_config 00:38:40.276 09:14:30 spdkcli_nvmf_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:38:40.276 09:14:30 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:40.276 09:14:30 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 32 512 Malloc1'\'' '\''Malloc1'\'' True 00:38:40.276 '\''/bdevs/malloc create 32 512 Malloc2'\'' '\''Malloc2'\'' True 00:38:40.276 '\''/bdevs/malloc create 32 512 Malloc3'\'' '\''Malloc3'\'' True 00:38:40.276 '\''/bdevs/malloc create 32 512 Malloc4'\'' '\''Malloc4'\'' True 00:38:40.276 '\''/bdevs/malloc create 32 512 Malloc5'\'' '\''Malloc5'\'' True 00:38:40.276 '\''/bdevs/malloc create 32 512 Malloc6'\'' '\''Malloc6'\'' True 00:38:40.276 '\''nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192'\'' '\'''\'' True 00:38:40.276 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1'\'' '\''Malloc3'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2'\'' '\''Malloc4'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:38:40.276 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2'\'' '\''Malloc2'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:38:40.276 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1'\'' '\''Malloc1'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True'\'' '\''Allow any host'\'' 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False'\'' '\''Allow any host'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4'\'' '\''127.0.0.1:4262'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5'\'' '\''Malloc5'\'' True 00:38:40.276 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6'\'' '\''Malloc6'\'' True 00:38:40.276 '\''/nvmf/referral create tcp 127.0.0.2 4030 IPv4'\'' 00:38:40.276 ' 00:38:42.819 [2024-11-06 09:14:32.739695] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:38:44.205 [2024-11-06 09:14:33.947580] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4260 *** 00:38:46.120 [2024-11-06 09:14:36.166323] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4261 *** 00:38:48.035 [2024-11-06 09:14:38.072018] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4262 *** 00:38:49.947 Executing command: ['/bdevs/malloc create 32 512 Malloc1', 'Malloc1', True] 00:38:49.947 Executing command: ['/bdevs/malloc create 32 512 Malloc2', 'Malloc2', True] 00:38:49.947 Executing command: ['/bdevs/malloc create 32 512 Malloc3', 'Malloc3', True] 00:38:49.947 Executing command: ['/bdevs/malloc create 32 512 Malloc4', 'Malloc4', True] 00:38:49.947 Executing command: ['/bdevs/malloc create 32 512 Malloc5', 'Malloc5', True] 00:38:49.947 Executing command: ['/bdevs/malloc create 32 512 Malloc6', 'Malloc6', True] 00:38:49.947 Executing command: ['nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192', '', True] 00:38:49.947 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode1', True] 00:38:49.947 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1', 'Malloc3', True] 00:38:49.947 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2', 'Malloc4', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:38:49.948 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2', 'Malloc2', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:38:49.948 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1', 'Malloc1', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1', 'nqn.2014-08.org.spdk:cnode1', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True', 'Allow any host', False] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False', 'Allow any host', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4', '127.0.0.1:4262', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5', 'Malloc5', True] 00:38:49.948 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6', 'Malloc6', True] 00:38:49.948 Executing command: ['/nvmf/referral create tcp 127.0.0.2 4030 IPv4', False] 00:38:49.948 09:14:39 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@66 -- # timing_exit spdkcli_create_nvmf_config 00:38:49.948 09:14:39 spdkcli_nvmf_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:38:49.948 09:14:39 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:49.948 09:14:39 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@68 -- # timing_enter spdkcli_check_match 00:38:49.948 09:14:39 spdkcli_nvmf_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:38:49.948 09:14:39 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:49.948 09:14:39 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@69 -- # check_match 00:38:49.948 09:14:39 spdkcli_nvmf_tcp -- spdkcli/common.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdkcli.py ll /nvmf 00:38:49.948 09:14:40 spdkcli_nvmf_tcp -- spdkcli/common.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/match/match /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test.match 00:38:50.208 09:14:40 spdkcli_nvmf_tcp -- spdkcli/common.sh@46 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test 00:38:50.208 09:14:40 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@70 -- # timing_exit spdkcli_check_match 00:38:50.208 09:14:40 spdkcli_nvmf_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:38:50.208 09:14:40 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:50.208 09:14:40 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@72 -- # timing_enter spdkcli_clear_nvmf_config 00:38:50.208 09:14:40 spdkcli_nvmf_tcp -- common/autotest_common.sh@724 -- # xtrace_disable 00:38:50.208 09:14:40 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:50.208 09:14:40 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1'\'' '\''Malloc3'\'' 00:38:50.208 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all'\'' '\''Malloc4'\'' 00:38:50.208 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:38:50.208 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' 00:38:50.208 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262'\'' '\''127.0.0.1:4262'\'' 00:38:50.208 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all'\'' '\''127.0.0.1:4261'\'' 00:38:50.208 '\''/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3'\'' '\''nqn.2014-08.org.spdk:cnode3'\'' 00:38:50.208 '\''/nvmf/subsystem delete_all'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:38:50.208 '\''/bdevs/malloc delete Malloc6'\'' '\''Malloc6'\'' 00:38:50.208 '\''/bdevs/malloc delete Malloc5'\'' '\''Malloc5'\'' 00:38:50.208 '\''/bdevs/malloc delete Malloc4'\'' '\''Malloc4'\'' 00:38:50.208 '\''/bdevs/malloc delete Malloc3'\'' '\''Malloc3'\'' 00:38:50.208 '\''/bdevs/malloc delete Malloc2'\'' '\''Malloc2'\'' 00:38:50.208 '\''/bdevs/malloc delete Malloc1'\'' '\''Malloc1'\'' 00:38:50.208 ' 00:38:55.493 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1', 'Malloc3', False] 00:38:55.493 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all', 'Malloc4', False] 00:38:55.493 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', False] 00:38:55.493 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all', 'nqn.2014-08.org.spdk:cnode1', False] 00:38:55.493 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262', '127.0.0.1:4262', False] 00:38:55.493 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all', '127.0.0.1:4261', False] 00:38:55.493 Executing command: ['/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3', 'nqn.2014-08.org.spdk:cnode3', False] 00:38:55.493 Executing command: ['/nvmf/subsystem delete_all', 'nqn.2014-08.org.spdk:cnode2', False] 00:38:55.493 Executing command: ['/bdevs/malloc delete Malloc6', 'Malloc6', False] 00:38:55.493 Executing command: ['/bdevs/malloc delete Malloc5', 'Malloc5', False] 00:38:55.493 Executing command: ['/bdevs/malloc delete Malloc4', 'Malloc4', False] 00:38:55.493 Executing command: ['/bdevs/malloc delete Malloc3', 'Malloc3', False] 00:38:55.493 Executing command: ['/bdevs/malloc delete Malloc2', 'Malloc2', False] 00:38:55.493 Executing command: ['/bdevs/malloc delete Malloc1', 'Malloc1', False] 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@88 -- # timing_exit spdkcli_clear_nvmf_config 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@730 -- # xtrace_disable 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@90 -- # killprocess 2178631 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@950 -- # '[' -z 2178631 ']' 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@954 -- # kill -0 2178631 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@955 -- # uname 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2178631 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2178631' 00:38:55.493 killing process with pid 2178631 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@969 -- # kill 2178631 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@974 -- # wait 2178631 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@1 -- # cleanup 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- spdkcli/common.sh@10 -- # '[' -n '' ']' 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- spdkcli/common.sh@13 -- # '[' -n 2178631 ']' 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- spdkcli/common.sh@14 -- # killprocess 2178631 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@950 -- # '[' -z 2178631 ']' 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@954 -- # kill -0 2178631 00:38:55.493 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (2178631) - No such process 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@977 -- # echo 'Process with pid 2178631 is not found' 00:38:55.493 Process with pid 2178631 is not found 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- spdkcli/common.sh@22 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_nvmf.test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:38:55.493 00:38:55.493 real 0m16.235s 00:38:55.493 user 0m33.623s 00:38:55.493 sys 0m0.701s 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@1126 -- # xtrace_disable 00:38:55.493 09:14:45 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:38:55.493 ************************************ 00:38:55.493 END TEST spdkcli_nvmf_tcp 00:38:55.493 ************************************ 00:38:55.493 09:14:45 -- spdk/autotest.sh@283 -- # run_test nvmf_identify_passthru /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:38:55.493 09:14:45 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:38:55.493 09:14:45 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:38:55.493 09:14:45 -- common/autotest_common.sh@10 -- # set +x 00:38:55.493 ************************************ 00:38:55.493 START TEST nvmf_identify_passthru 00:38:55.493 ************************************ 00:38:55.493 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:38:55.493 * Looking for test storage... 00:38:55.493 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:38:55.493 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:38:55.755 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@1689 -- # lcov --version 00:38:55.755 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:38:55.755 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@333 -- # local ver1 ver1_l 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@334 -- # local ver2 ver2_l 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@336 -- # IFS=.-: 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@336 -- # read -ra ver1 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@337 -- # IFS=.-: 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@337 -- # read -ra ver2 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@338 -- # local 'op=<' 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@340 -- # ver1_l=2 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@341 -- # ver2_l=1 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@344 -- # case "$op" in 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@345 -- # : 1 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@364 -- # (( v = 0 )) 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@365 -- # decimal 1 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@353 -- # local d=1 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@355 -- # echo 1 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@365 -- # ver1[v]=1 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@366 -- # decimal 2 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@353 -- # local d=2 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@355 -- # echo 2 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@366 -- # ver2[v]=2 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:38:55.755 09:14:45 nvmf_identify_passthru -- scripts/common.sh@368 -- # return 0 00:38:55.755 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:38:55.755 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:38:55.755 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.755 --rc genhtml_branch_coverage=1 00:38:55.755 --rc genhtml_function_coverage=1 00:38:55.755 --rc genhtml_legend=1 00:38:55.755 --rc geninfo_all_blocks=1 00:38:55.755 --rc geninfo_unexecuted_blocks=1 00:38:55.755 00:38:55.755 ' 00:38:55.755 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:38:55.755 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.755 --rc genhtml_branch_coverage=1 00:38:55.755 --rc genhtml_function_coverage=1 00:38:55.755 --rc genhtml_legend=1 00:38:55.755 --rc geninfo_all_blocks=1 00:38:55.755 --rc geninfo_unexecuted_blocks=1 00:38:55.755 00:38:55.755 ' 00:38:55.755 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:38:55.755 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.755 --rc genhtml_branch_coverage=1 00:38:55.755 --rc genhtml_function_coverage=1 00:38:55.756 --rc genhtml_legend=1 00:38:55.756 --rc geninfo_all_blocks=1 00:38:55.756 --rc geninfo_unexecuted_blocks=1 00:38:55.756 00:38:55.756 ' 00:38:55.756 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:38:55.756 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:38:55.756 --rc genhtml_branch_coverage=1 00:38:55.756 --rc genhtml_function_coverage=1 00:38:55.756 --rc genhtml_legend=1 00:38:55.756 --rc geninfo_all_blocks=1 00:38:55.756 --rc geninfo_unexecuted_blocks=1 00:38:55.756 00:38:55.756 ' 00:38:55.756 09:14:45 nvmf_identify_passthru -- target/identify_passthru.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@7 -- # uname -s 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:55.756 09:14:45 nvmf_identify_passthru -- scripts/common.sh@15 -- # shopt -s extglob 00:38:55.756 09:14:45 nvmf_identify_passthru -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:38:55.756 09:14:45 nvmf_identify_passthru -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:55.756 09:14:45 nvmf_identify_passthru -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:55.756 09:14:45 nvmf_identify_passthru -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.756 09:14:45 nvmf_identify_passthru -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.756 09:14:45 nvmf_identify_passthru -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.756 09:14:45 nvmf_identify_passthru -- paths/export.sh@5 -- # export PATH 00:38:55.756 09:14:45 nvmf_identify_passthru -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@51 -- # : 0 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:38:55.756 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@55 -- # have_pci_nics=0 00:38:55.756 09:14:45 nvmf_identify_passthru -- target/identify_passthru.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:38:55.756 09:14:45 nvmf_identify_passthru -- scripts/common.sh@15 -- # shopt -s extglob 00:38:55.756 09:14:45 nvmf_identify_passthru -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:38:55.756 09:14:45 nvmf_identify_passthru -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:38:55.756 09:14:45 nvmf_identify_passthru -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:38:55.756 09:14:45 nvmf_identify_passthru -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.756 09:14:45 nvmf_identify_passthru -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.756 09:14:45 nvmf_identify_passthru -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.756 09:14:45 nvmf_identify_passthru -- paths/export.sh@5 -- # export PATH 00:38:55.756 09:14:45 nvmf_identify_passthru -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:38:55.756 09:14:45 nvmf_identify_passthru -- target/identify_passthru.sh@12 -- # nvmftestinit 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@474 -- # prepare_net_devs 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@436 -- # local -g is_hw=no 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@438 -- # remove_spdk_ns 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:38:55.756 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:38:55.756 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:38:55.756 09:14:45 nvmf_identify_passthru -- nvmf/common.sh@309 -- # xtrace_disable 00:38:55.756 09:14:45 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@315 -- # pci_devs=() 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@315 -- # local -a pci_devs 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@316 -- # pci_net_devs=() 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@317 -- # pci_drivers=() 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@317 -- # local -A pci_drivers 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@319 -- # net_devs=() 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@319 -- # local -ga net_devs 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@320 -- # e810=() 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@320 -- # local -ga e810 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@321 -- # x722=() 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@321 -- # local -ga x722 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@322 -- # mlx=() 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@322 -- # local -ga mlx 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:39:03.895 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:39:03.895 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:39:03.895 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@416 -- # [[ up == up ]] 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:39:03.896 Found net devices under 0000:4b:00.0: cvl_0_0 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@416 -- # [[ up == up ]] 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:39:03.896 Found net devices under 0000:4b:00.1: cvl_0_1 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@440 -- # is_hw=yes 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:39:03.896 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:39:03.896 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.607 ms 00:39:03.896 00:39:03.896 --- 10.0.0.2 ping statistics --- 00:39:03.896 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:39:03.896 rtt min/avg/max/mdev = 0.607/0.607/0.607/0.000 ms 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:39:03.896 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:39:03.896 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.292 ms 00:39:03.896 00:39:03.896 --- 10.0.0.1 ping statistics --- 00:39:03.896 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:39:03.896 rtt min/avg/max/mdev = 0.292/0.292/0.292/0.000 ms 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@448 -- # return 0 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@476 -- # '[' '' == iso ']' 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:39:03.896 09:14:52 nvmf_identify_passthru -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@14 -- # timing_enter nvme_identify 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@724 -- # xtrace_disable 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@16 -- # get_first_nvme_bdf 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1505 -- # bdfs=() 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1505 -- # local bdfs 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1506 -- # bdfs=($(get_nvme_bdfs)) 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1506 -- # get_nvme_bdfs 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1494 -- # bdfs=() 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1494 -- # local bdfs 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1495 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1495 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1495 -- # jq -r '.config[].params.traddr' 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1496 -- # (( 1 == 0 )) 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1500 -- # printf '%s\n' 0000:65:00.0 00:39:03.896 09:14:53 nvmf_identify_passthru -- common/autotest_common.sh@1508 -- # echo 0000:65:00.0 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@16 -- # bdf=0000:65:00.0 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@17 -- # '[' -z 0000:65:00.0 ']' 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:65:00.0' -i 0 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # grep 'Serial Number:' 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # awk '{print $3}' 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # nvme_serial_number=S64GNE0R605487 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:65:00.0' -i 0 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # grep 'Model Number:' 00:39:03.896 09:14:53 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # awk '{print $3}' 00:39:04.157 09:14:54 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # nvme_model_number=SAMSUNG 00:39:04.157 09:14:54 nvmf_identify_passthru -- target/identify_passthru.sh@26 -- # timing_exit nvme_identify 00:39:04.157 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@730 -- # xtrace_disable 00:39:04.157 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:04.157 09:14:54 nvmf_identify_passthru -- target/identify_passthru.sh@28 -- # timing_enter start_nvmf_tgt 00:39:04.157 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@724 -- # xtrace_disable 00:39:04.157 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:04.157 09:14:54 nvmf_identify_passthru -- target/identify_passthru.sh@31 -- # nvmfpid=2185682 00:39:04.157 09:14:54 nvmf_identify_passthru -- target/identify_passthru.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:39:04.157 09:14:54 nvmf_identify_passthru -- target/identify_passthru.sh@30 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:39:04.157 09:14:54 nvmf_identify_passthru -- target/identify_passthru.sh@35 -- # waitforlisten 2185682 00:39:04.157 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@831 -- # '[' -z 2185682 ']' 00:39:04.157 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:39:04.157 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@836 -- # local max_retries=100 00:39:04.157 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:39:04.157 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:39:04.157 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@840 -- # xtrace_disable 00:39:04.157 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:04.157 [2024-11-06 09:14:54.179826] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:39:04.157 [2024-11-06 09:14:54.179878] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:39:04.157 [2024-11-06 09:14:54.256452] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:39:04.417 [2024-11-06 09:14:54.292693] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:39:04.417 [2024-11-06 09:14:54.292728] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:39:04.417 [2024-11-06 09:14:54.292737] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:39:04.417 [2024-11-06 09:14:54.292744] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:39:04.417 [2024-11-06 09:14:54.292756] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:39:04.417 [2024-11-06 09:14:54.294473] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:39:04.417 [2024-11-06 09:14:54.294585] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:39:04.417 [2024-11-06 09:14:54.294739] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:39:04.417 [2024-11-06 09:14:54.294741] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:39:04.987 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:39:04.987 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@864 -- # return 0 00:39:04.987 09:14:54 nvmf_identify_passthru -- target/identify_passthru.sh@36 -- # rpc_cmd -v nvmf_set_config --passthru-identify-ctrlr 00:39:04.987 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:04.987 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:04.987 INFO: Log level set to 20 00:39:04.987 INFO: Requests: 00:39:04.987 { 00:39:04.987 "jsonrpc": "2.0", 00:39:04.988 "method": "nvmf_set_config", 00:39:04.988 "id": 1, 00:39:04.988 "params": { 00:39:04.988 "admin_cmd_passthru": { 00:39:04.988 "identify_ctrlr": true 00:39:04.988 } 00:39:04.988 } 00:39:04.988 } 00:39:04.988 00:39:04.988 INFO: response: 00:39:04.988 { 00:39:04.988 "jsonrpc": "2.0", 00:39:04.988 "id": 1, 00:39:04.988 "result": true 00:39:04.988 } 00:39:04.988 00:39:04.988 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:04.988 09:14:54 nvmf_identify_passthru -- target/identify_passthru.sh@37 -- # rpc_cmd -v framework_start_init 00:39:04.988 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:04.988 09:14:54 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:04.988 INFO: Setting log level to 20 00:39:04.988 INFO: Setting log level to 20 00:39:04.988 INFO: Log level set to 20 00:39:04.988 INFO: Log level set to 20 00:39:04.988 INFO: Requests: 00:39:04.988 { 00:39:04.988 "jsonrpc": "2.0", 00:39:04.988 "method": "framework_start_init", 00:39:04.988 "id": 1 00:39:04.988 } 00:39:04.988 00:39:04.988 INFO: Requests: 00:39:04.988 { 00:39:04.988 "jsonrpc": "2.0", 00:39:04.988 "method": "framework_start_init", 00:39:04.988 "id": 1 00:39:04.988 } 00:39:04.988 00:39:04.988 [2024-11-06 09:14:55.052115] nvmf_tgt.c: 462:nvmf_tgt_advance_state: *NOTICE*: Custom identify ctrlr handler enabled 00:39:04.988 INFO: response: 00:39:04.988 { 00:39:04.988 "jsonrpc": "2.0", 00:39:04.988 "id": 1, 00:39:04.988 "result": true 00:39:04.988 } 00:39:04.988 00:39:04.988 INFO: response: 00:39:04.988 { 00:39:04.988 "jsonrpc": "2.0", 00:39:04.988 "id": 1, 00:39:04.988 "result": true 00:39:04.988 } 00:39:04.988 00:39:04.988 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:04.988 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@38 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:39:04.988 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:04.988 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:04.988 INFO: Setting log level to 40 00:39:04.988 INFO: Setting log level to 40 00:39:04.988 INFO: Setting log level to 40 00:39:04.988 [2024-11-06 09:14:55.065436] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:39:04.988 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:04.988 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@39 -- # timing_exit start_nvmf_tgt 00:39:04.988 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@730 -- # xtrace_disable 00:39:04.988 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:05.249 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@41 -- # rpc_cmd bdev_nvme_attach_controller -b Nvme0 -t PCIe -a 0000:65:00.0 00:39:05.249 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:05.249 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:05.510 Nvme0n1 00:39:05.510 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:05.510 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@42 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 1 00:39:05.510 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:05.510 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:05.510 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:05.510 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@43 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:39:05.511 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:05.511 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:05.511 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:05.511 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@44 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:39:05.511 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:05.511 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:05.511 [2024-11-06 09:14:55.465071] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:39:05.511 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:05.511 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@46 -- # rpc_cmd nvmf_get_subsystems 00:39:05.511 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:05.511 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:05.511 [ 00:39:05.511 { 00:39:05.511 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:39:05.511 "subtype": "Discovery", 00:39:05.511 "listen_addresses": [], 00:39:05.511 "allow_any_host": true, 00:39:05.511 "hosts": [] 00:39:05.511 }, 00:39:05.511 { 00:39:05.511 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:39:05.511 "subtype": "NVMe", 00:39:05.511 "listen_addresses": [ 00:39:05.511 { 00:39:05.511 "trtype": "TCP", 00:39:05.511 "adrfam": "IPv4", 00:39:05.511 "traddr": "10.0.0.2", 00:39:05.511 "trsvcid": "4420" 00:39:05.511 } 00:39:05.511 ], 00:39:05.511 "allow_any_host": true, 00:39:05.511 "hosts": [], 00:39:05.511 "serial_number": "SPDK00000000000001", 00:39:05.511 "model_number": "SPDK bdev Controller", 00:39:05.511 "max_namespaces": 1, 00:39:05.511 "min_cntlid": 1, 00:39:05.511 "max_cntlid": 65519, 00:39:05.511 "namespaces": [ 00:39:05.511 { 00:39:05.511 "nsid": 1, 00:39:05.511 "bdev_name": "Nvme0n1", 00:39:05.511 "name": "Nvme0n1", 00:39:05.511 "nguid": "36344730526054870025384500000044", 00:39:05.511 "uuid": "36344730-5260-5487-0025-384500000044" 00:39:05.511 } 00:39:05.511 ] 00:39:05.511 } 00:39:05.511 ] 00:39:05.511 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:05.511 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:39:05.511 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # grep 'Serial Number:' 00:39:05.511 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # awk '{print $3}' 00:39:05.772 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # nvmf_serial_number=S64GNE0R605487 00:39:05.772 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:39:05.772 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # grep 'Model Number:' 00:39:05.772 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # awk '{print $3}' 00:39:05.772 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # nvmf_model_number=SAMSUNG 00:39:05.772 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@63 -- # '[' S64GNE0R605487 '!=' S64GNE0R605487 ']' 00:39:05.772 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@68 -- # '[' SAMSUNG '!=' SAMSUNG ']' 00:39:05.772 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@73 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:39:05.772 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:05.772 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:05.772 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:05.772 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@75 -- # trap - SIGINT SIGTERM EXIT 00:39:05.772 09:14:55 nvmf_identify_passthru -- target/identify_passthru.sh@77 -- # nvmftestfini 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@514 -- # nvmfcleanup 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@121 -- # sync 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@124 -- # set +e 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@125 -- # for i in {1..20} 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:39:05.772 rmmod nvme_tcp 00:39:05.772 rmmod nvme_fabrics 00:39:05.772 rmmod nvme_keyring 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@128 -- # set -e 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@129 -- # return 0 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@515 -- # '[' -n 2185682 ']' 00:39:05.772 09:14:55 nvmf_identify_passthru -- nvmf/common.sh@516 -- # killprocess 2185682 00:39:05.772 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@950 -- # '[' -z 2185682 ']' 00:39:05.772 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@954 -- # kill -0 2185682 00:39:05.772 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@955 -- # uname 00:39:05.772 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:39:05.772 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2185682 00:39:06.033 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:39:06.033 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:39:06.033 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2185682' 00:39:06.033 killing process with pid 2185682 00:39:06.033 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@969 -- # kill 2185682 00:39:06.033 09:14:55 nvmf_identify_passthru -- common/autotest_common.sh@974 -- # wait 2185682 00:39:06.294 09:14:56 nvmf_identify_passthru -- nvmf/common.sh@518 -- # '[' '' == iso ']' 00:39:06.294 09:14:56 nvmf_identify_passthru -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:39:06.294 09:14:56 nvmf_identify_passthru -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:39:06.294 09:14:56 nvmf_identify_passthru -- nvmf/common.sh@297 -- # iptr 00:39:06.294 09:14:56 nvmf_identify_passthru -- nvmf/common.sh@789 -- # iptables-save 00:39:06.294 09:14:56 nvmf_identify_passthru -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:39:06.294 09:14:56 nvmf_identify_passthru -- nvmf/common.sh@789 -- # iptables-restore 00:39:06.294 09:14:56 nvmf_identify_passthru -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:39:06.294 09:14:56 nvmf_identify_passthru -- nvmf/common.sh@302 -- # remove_spdk_ns 00:39:06.294 09:14:56 nvmf_identify_passthru -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:39:06.294 09:14:56 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:39:06.294 09:14:56 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:39:08.209 09:14:58 nvmf_identify_passthru -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:39:08.209 00:39:08.209 real 0m12.760s 00:39:08.209 user 0m9.694s 00:39:08.209 sys 0m6.536s 00:39:08.209 09:14:58 nvmf_identify_passthru -- common/autotest_common.sh@1126 -- # xtrace_disable 00:39:08.209 09:14:58 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:39:08.209 ************************************ 00:39:08.209 END TEST nvmf_identify_passthru 00:39:08.209 ************************************ 00:39:08.209 09:14:58 -- spdk/autotest.sh@285 -- # run_test nvmf_dif /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:39:08.209 09:14:58 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:39:08.209 09:14:58 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:39:08.209 09:14:58 -- common/autotest_common.sh@10 -- # set +x 00:39:08.491 ************************************ 00:39:08.491 START TEST nvmf_dif 00:39:08.491 ************************************ 00:39:08.491 09:14:58 nvmf_dif -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:39:08.491 * Looking for test storage... 00:39:08.491 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:39:08.491 09:14:58 nvmf_dif -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:39:08.491 09:14:58 nvmf_dif -- common/autotest_common.sh@1689 -- # lcov --version 00:39:08.491 09:14:58 nvmf_dif -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:39:08.491 09:14:58 nvmf_dif -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@333 -- # local ver1 ver1_l 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@334 -- # local ver2 ver2_l 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@336 -- # IFS=.-: 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@336 -- # read -ra ver1 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@337 -- # IFS=.-: 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@337 -- # read -ra ver2 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@338 -- # local 'op=<' 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@340 -- # ver1_l=2 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@341 -- # ver2_l=1 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@344 -- # case "$op" in 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@345 -- # : 1 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@364 -- # (( v = 0 )) 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@365 -- # decimal 1 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@353 -- # local d=1 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@355 -- # echo 1 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@365 -- # ver1[v]=1 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@366 -- # decimal 2 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@353 -- # local d=2 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@355 -- # echo 2 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@366 -- # ver2[v]=2 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:39:08.491 09:14:58 nvmf_dif -- scripts/common.sh@368 -- # return 0 00:39:08.491 09:14:58 nvmf_dif -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:39:08.491 09:14:58 nvmf_dif -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:39:08.491 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:08.491 --rc genhtml_branch_coverage=1 00:39:08.491 --rc genhtml_function_coverage=1 00:39:08.491 --rc genhtml_legend=1 00:39:08.491 --rc geninfo_all_blocks=1 00:39:08.491 --rc geninfo_unexecuted_blocks=1 00:39:08.491 00:39:08.491 ' 00:39:08.491 09:14:58 nvmf_dif -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:39:08.491 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:08.491 --rc genhtml_branch_coverage=1 00:39:08.491 --rc genhtml_function_coverage=1 00:39:08.491 --rc genhtml_legend=1 00:39:08.491 --rc geninfo_all_blocks=1 00:39:08.491 --rc geninfo_unexecuted_blocks=1 00:39:08.491 00:39:08.491 ' 00:39:08.491 09:14:58 nvmf_dif -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:39:08.491 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:08.491 --rc genhtml_branch_coverage=1 00:39:08.491 --rc genhtml_function_coverage=1 00:39:08.491 --rc genhtml_legend=1 00:39:08.491 --rc geninfo_all_blocks=1 00:39:08.491 --rc geninfo_unexecuted_blocks=1 00:39:08.491 00:39:08.491 ' 00:39:08.491 09:14:58 nvmf_dif -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:39:08.492 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:39:08.492 --rc genhtml_branch_coverage=1 00:39:08.492 --rc genhtml_function_coverage=1 00:39:08.492 --rc genhtml_legend=1 00:39:08.492 --rc geninfo_all_blocks=1 00:39:08.492 --rc geninfo_unexecuted_blocks=1 00:39:08.492 00:39:08.492 ' 00:39:08.492 09:14:58 nvmf_dif -- target/dif.sh@13 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@7 -- # uname -s 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:39:08.492 09:14:58 nvmf_dif -- scripts/common.sh@15 -- # shopt -s extglob 00:39:08.492 09:14:58 nvmf_dif -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:39:08.492 09:14:58 nvmf_dif -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:39:08.492 09:14:58 nvmf_dif -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:39:08.492 09:14:58 nvmf_dif -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:08.492 09:14:58 nvmf_dif -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:08.492 09:14:58 nvmf_dif -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:08.492 09:14:58 nvmf_dif -- paths/export.sh@5 -- # export PATH 00:39:08.492 09:14:58 nvmf_dif -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@51 -- # : 0 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:39:08.492 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@55 -- # have_pci_nics=0 00:39:08.492 09:14:58 nvmf_dif -- target/dif.sh@15 -- # NULL_META=16 00:39:08.492 09:14:58 nvmf_dif -- target/dif.sh@15 -- # NULL_BLOCK_SIZE=512 00:39:08.492 09:14:58 nvmf_dif -- target/dif.sh@15 -- # NULL_SIZE=64 00:39:08.492 09:14:58 nvmf_dif -- target/dif.sh@15 -- # NULL_DIF=1 00:39:08.492 09:14:58 nvmf_dif -- target/dif.sh@135 -- # nvmftestinit 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@474 -- # prepare_net_devs 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@436 -- # local -g is_hw=no 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@438 -- # remove_spdk_ns 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:39:08.492 09:14:58 nvmf_dif -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:39:08.492 09:14:58 nvmf_dif -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:39:08.492 09:14:58 nvmf_dif -- nvmf/common.sh@309 -- # xtrace_disable 00:39:08.492 09:14:58 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@315 -- # pci_devs=() 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@315 -- # local -a pci_devs 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@316 -- # pci_net_devs=() 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@317 -- # pci_drivers=() 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@317 -- # local -A pci_drivers 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@319 -- # net_devs=() 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@319 -- # local -ga net_devs 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@320 -- # e810=() 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@320 -- # local -ga e810 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@321 -- # x722=() 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@321 -- # local -ga x722 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@322 -- # mlx=() 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@322 -- # local -ga mlx 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:39:15.124 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:39:15.124 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@416 -- # [[ up == up ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:39:15.124 Found net devices under 0000:4b:00.0: cvl_0_0 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@416 -- # [[ up == up ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:39:15.124 Found net devices under 0000:4b:00.1: cvl_0_1 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@440 -- # is_hw=yes 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:39:15.124 09:15:05 nvmf_dif -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:39:15.385 09:15:05 nvmf_dif -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:39:15.385 09:15:05 nvmf_dif -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:39:15.385 09:15:05 nvmf_dif -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:39:15.385 09:15:05 nvmf_dif -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:39:15.385 09:15:05 nvmf_dif -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:39:15.385 09:15:05 nvmf_dif -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:39:15.385 09:15:05 nvmf_dif -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:39:15.385 09:15:05 nvmf_dif -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:39:15.385 09:15:05 nvmf_dif -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:39:15.385 09:15:05 nvmf_dif -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:39:15.647 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:39:15.647 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.762 ms 00:39:15.647 00:39:15.647 --- 10.0.0.2 ping statistics --- 00:39:15.647 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:39:15.647 rtt min/avg/max/mdev = 0.762/0.762/0.762/0.000 ms 00:39:15.647 09:15:05 nvmf_dif -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:39:15.647 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:39:15.647 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.280 ms 00:39:15.647 00:39:15.647 --- 10.0.0.1 ping statistics --- 00:39:15.647 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:39:15.647 rtt min/avg/max/mdev = 0.280/0.280/0.280/0.000 ms 00:39:15.647 09:15:05 nvmf_dif -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:39:15.647 09:15:05 nvmf_dif -- nvmf/common.sh@448 -- # return 0 00:39:15.647 09:15:05 nvmf_dif -- nvmf/common.sh@476 -- # '[' iso == iso ']' 00:39:15.647 09:15:05 nvmf_dif -- nvmf/common.sh@477 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:39:18.952 0000:80:01.6 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:80:01.7 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:80:01.4 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:80:01.5 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:80:01.2 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:80:01.3 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:80:01.0 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:80:01.1 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:00:01.6 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:65:00.0 (144d a80a): Already using the vfio-pci driver 00:39:18.952 0000:00:01.7 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:00:01.4 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:00:01.5 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:00:01.2 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:00:01.3 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:00:01.0 (8086 0b00): Already using the vfio-pci driver 00:39:18.952 0000:00:01.1 (8086 0b00): Already using the vfio-pci driver 00:39:19.213 09:15:09 nvmf_dif -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:39:19.213 09:15:09 nvmf_dif -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:39:19.214 09:15:09 nvmf_dif -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:39:19.214 09:15:09 nvmf_dif -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:39:19.214 09:15:09 nvmf_dif -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:39:19.214 09:15:09 nvmf_dif -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:39:19.214 09:15:09 nvmf_dif -- target/dif.sh@136 -- # NVMF_TRANSPORT_OPTS+=' --dif-insert-or-strip' 00:39:19.214 09:15:09 nvmf_dif -- target/dif.sh@137 -- # nvmfappstart 00:39:19.214 09:15:09 nvmf_dif -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:39:19.214 09:15:09 nvmf_dif -- common/autotest_common.sh@724 -- # xtrace_disable 00:39:19.214 09:15:09 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:39:19.214 09:15:09 nvmf_dif -- nvmf/common.sh@507 -- # nvmfpid=2191868 00:39:19.214 09:15:09 nvmf_dif -- nvmf/common.sh@508 -- # waitforlisten 2191868 00:39:19.214 09:15:09 nvmf_dif -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:39:19.214 09:15:09 nvmf_dif -- common/autotest_common.sh@831 -- # '[' -z 2191868 ']' 00:39:19.214 09:15:09 nvmf_dif -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:39:19.214 09:15:09 nvmf_dif -- common/autotest_common.sh@836 -- # local max_retries=100 00:39:19.214 09:15:09 nvmf_dif -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:39:19.214 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:39:19.214 09:15:09 nvmf_dif -- common/autotest_common.sh@840 -- # xtrace_disable 00:39:19.214 09:15:09 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:39:19.214 [2024-11-06 09:15:09.293695] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:39:19.214 [2024-11-06 09:15:09.293744] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:39:19.475 [2024-11-06 09:15:09.369430] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:39:19.475 [2024-11-06 09:15:09.404173] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:39:19.475 [2024-11-06 09:15:09.404205] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:39:19.475 [2024-11-06 09:15:09.404213] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:39:19.475 [2024-11-06 09:15:09.404220] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:39:19.475 [2024-11-06 09:15:09.404226] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:39:19.475 [2024-11-06 09:15:09.404769] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:39:20.046 09:15:10 nvmf_dif -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:39:20.046 09:15:10 nvmf_dif -- common/autotest_common.sh@864 -- # return 0 00:39:20.047 09:15:10 nvmf_dif -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:39:20.047 09:15:10 nvmf_dif -- common/autotest_common.sh@730 -- # xtrace_disable 00:39:20.047 09:15:10 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:39:20.047 09:15:10 nvmf_dif -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:39:20.047 09:15:10 nvmf_dif -- target/dif.sh@139 -- # create_transport 00:39:20.047 09:15:10 nvmf_dif -- target/dif.sh@50 -- # rpc_cmd nvmf_create_transport -t tcp -o --dif-insert-or-strip 00:39:20.047 09:15:10 nvmf_dif -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:20.047 09:15:10 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:39:20.047 [2024-11-06 09:15:10.121956] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:39:20.047 09:15:10 nvmf_dif -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:20.047 09:15:10 nvmf_dif -- target/dif.sh@141 -- # run_test fio_dif_1_default fio_dif_1 00:39:20.047 09:15:10 nvmf_dif -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:39:20.047 09:15:10 nvmf_dif -- common/autotest_common.sh@1107 -- # xtrace_disable 00:39:20.047 09:15:10 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:39:20.309 ************************************ 00:39:20.309 START TEST fio_dif_1_default 00:39:20.309 ************************************ 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1125 -- # fio_dif_1 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@86 -- # create_subsystems 0 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@28 -- # local sub 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@30 -- # for sub in "$@" 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@31 -- # create_subsystem 0 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@18 -- # local sub_id=0 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:39:20.309 bdev_null0 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:39:20.309 [2024-11-06 09:15:10.210344] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@87 -- # fio /dev/fd/62 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@87 -- # create_json_sub_conf 0 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@558 -- # config=() 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@558 -- # local subsystem config 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:39:20.309 { 00:39:20.309 "params": { 00:39:20.309 "name": "Nvme$subsystem", 00:39:20.309 "trtype": "$TEST_TRANSPORT", 00:39:20.309 "traddr": "$NVMF_FIRST_TARGET_IP", 00:39:20.309 "adrfam": "ipv4", 00:39:20.309 "trsvcid": "$NVMF_PORT", 00:39:20.309 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:39:20.309 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:39:20.309 "hdgst": ${hdgst:-false}, 00:39:20.309 "ddgst": ${ddgst:-false} 00:39:20.309 }, 00:39:20.309 "method": "bdev_nvme_attach_controller" 00:39:20.309 } 00:39:20.309 EOF 00:39:20.309 )") 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@82 -- # gen_fio_conf 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@54 -- # local file 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1339 -- # local sanitizers 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@56 -- # cat 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1341 -- # shift 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1343 -- # local asan_lib= 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@580 -- # cat 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@72 -- # (( file = 1 )) 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # grep libasan 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- target/dif.sh@72 -- # (( file <= files )) 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@582 -- # jq . 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@583 -- # IFS=, 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:39:20.309 "params": { 00:39:20.309 "name": "Nvme0", 00:39:20.309 "trtype": "tcp", 00:39:20.309 "traddr": "10.0.0.2", 00:39:20.309 "adrfam": "ipv4", 00:39:20.309 "trsvcid": "4420", 00:39:20.309 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:39:20.309 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:39:20.309 "hdgst": false, 00:39:20.309 "ddgst": false 00:39:20.309 }, 00:39:20.309 "method": "bdev_nvme_attach_controller" 00:39:20.309 }' 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # asan_lib= 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # asan_lib= 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:39:20.309 09:15:10 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:20.570 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:39:20.570 fio-3.35 00:39:20.570 Starting 1 thread 00:39:32.809 00:39:32.809 filename0: (groupid=0, jobs=1): err= 0: pid=2192652: Wed Nov 6 09:15:21 2024 00:39:32.809 read: IOPS=188, BW=756KiB/s (774kB/s)(7568KiB/10016msec) 00:39:32.809 slat (nsec): min=5431, max=32049, avg=6151.83, stdev=1345.03 00:39:32.809 clat (usec): min=587, max=43006, avg=21157.90, stdev=20184.51 00:39:32.809 lat (usec): min=595, max=43015, avg=21164.05, stdev=20184.50 00:39:32.809 clat percentiles (usec): 00:39:32.809 | 1.00th=[ 873], 5.00th=[ 906], 10.00th=[ 914], 20.00th=[ 938], 00:39:32.809 | 30.00th=[ 947], 40.00th=[ 963], 50.00th=[41157], 60.00th=[41157], 00:39:32.809 | 70.00th=[41157], 80.00th=[41157], 90.00th=[42206], 95.00th=[42206], 00:39:32.809 | 99.00th=[42730], 99.50th=[42730], 99.90th=[43254], 99.95th=[43254], 00:39:32.809 | 99.99th=[43254] 00:39:32.809 bw ( KiB/s): min= 704, max= 768, per=99.92%, avg=755.20, stdev=26.27, samples=20 00:39:32.809 iops : min= 176, max= 192, avg=188.70, stdev= 6.53, samples=20 00:39:32.809 lat (usec) : 750=0.63%, 1000=48.10% 00:39:32.809 lat (msec) : 2=1.16%, 50=50.11% 00:39:32.809 cpu : usr=93.19%, sys=6.60%, ctx=14, majf=0, minf=220 00:39:32.809 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:39:32.809 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:32.809 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:32.809 issued rwts: total=1892,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:39:32.809 latency : target=0, window=0, percentile=100.00%, depth=4 00:39:32.809 00:39:32.809 Run status group 0 (all jobs): 00:39:32.809 READ: bw=756KiB/s (774kB/s), 756KiB/s-756KiB/s (774kB/s-774kB/s), io=7568KiB (7750kB), run=10016-10016msec 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- target/dif.sh@88 -- # destroy_subsystems 0 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- target/dif.sh@43 -- # local sub 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- target/dif.sh@45 -- # for sub in "$@" 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- target/dif.sh@46 -- # destroy_subsystem 0 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- target/dif.sh@36 -- # local sub_id=0 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:32.809 00:39:32.809 real 0m11.271s 00:39:32.809 user 0m23.626s 00:39:32.809 sys 0m0.977s 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1126 -- # xtrace_disable 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:39:32.809 ************************************ 00:39:32.809 END TEST fio_dif_1_default 00:39:32.809 ************************************ 00:39:32.809 09:15:21 nvmf_dif -- target/dif.sh@142 -- # run_test fio_dif_1_multi_subsystems fio_dif_1_multi_subsystems 00:39:32.809 09:15:21 nvmf_dif -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:39:32.809 09:15:21 nvmf_dif -- common/autotest_common.sh@1107 -- # xtrace_disable 00:39:32.809 09:15:21 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:39:32.809 ************************************ 00:39:32.809 START TEST fio_dif_1_multi_subsystems 00:39:32.809 ************************************ 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1125 -- # fio_dif_1_multi_subsystems 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@92 -- # local files=1 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@94 -- # create_subsystems 0 1 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@28 -- # local sub 00:39:32.809 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@30 -- # for sub in "$@" 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@31 -- # create_subsystem 0 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@18 -- # local sub_id=0 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:32.810 bdev_null0 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:32.810 [2024-11-06 09:15:21.558813] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@30 -- # for sub in "$@" 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@31 -- # create_subsystem 1 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@18 -- # local sub_id=1 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:32.810 bdev_null1 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@95 -- # fio /dev/fd/62 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@95 -- # create_json_sub_conf 0 1 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@558 -- # config=() 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@558 -- # local subsystem config 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:39:32.810 { 00:39:32.810 "params": { 00:39:32.810 "name": "Nvme$subsystem", 00:39:32.810 "trtype": "$TEST_TRANSPORT", 00:39:32.810 "traddr": "$NVMF_FIRST_TARGET_IP", 00:39:32.810 "adrfam": "ipv4", 00:39:32.810 "trsvcid": "$NVMF_PORT", 00:39:32.810 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:39:32.810 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:39:32.810 "hdgst": ${hdgst:-false}, 00:39:32.810 "ddgst": ${ddgst:-false} 00:39:32.810 }, 00:39:32.810 "method": "bdev_nvme_attach_controller" 00:39:32.810 } 00:39:32.810 EOF 00:39:32.810 )") 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@82 -- # gen_fio_conf 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@54 -- # local file 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1339 -- # local sanitizers 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@56 -- # cat 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1341 -- # shift 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1343 -- # local asan_lib= 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@580 -- # cat 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file = 1 )) 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # grep libasan 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file <= files )) 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@73 -- # cat 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:39:32.810 { 00:39:32.810 "params": { 00:39:32.810 "name": "Nvme$subsystem", 00:39:32.810 "trtype": "$TEST_TRANSPORT", 00:39:32.810 "traddr": "$NVMF_FIRST_TARGET_IP", 00:39:32.810 "adrfam": "ipv4", 00:39:32.810 "trsvcid": "$NVMF_PORT", 00:39:32.810 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:39:32.810 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:39:32.810 "hdgst": ${hdgst:-false}, 00:39:32.810 "ddgst": ${ddgst:-false} 00:39:32.810 }, 00:39:32.810 "method": "bdev_nvme_attach_controller" 00:39:32.810 } 00:39:32.810 EOF 00:39:32.810 )") 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file++ )) 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file <= files )) 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@580 -- # cat 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@582 -- # jq . 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@583 -- # IFS=, 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:39:32.810 "params": { 00:39:32.810 "name": "Nvme0", 00:39:32.810 "trtype": "tcp", 00:39:32.810 "traddr": "10.0.0.2", 00:39:32.810 "adrfam": "ipv4", 00:39:32.810 "trsvcid": "4420", 00:39:32.810 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:39:32.810 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:39:32.810 "hdgst": false, 00:39:32.810 "ddgst": false 00:39:32.810 }, 00:39:32.810 "method": "bdev_nvme_attach_controller" 00:39:32.810 },{ 00:39:32.810 "params": { 00:39:32.810 "name": "Nvme1", 00:39:32.810 "trtype": "tcp", 00:39:32.810 "traddr": "10.0.0.2", 00:39:32.810 "adrfam": "ipv4", 00:39:32.810 "trsvcid": "4420", 00:39:32.810 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:39:32.810 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:39:32.810 "hdgst": false, 00:39:32.810 "ddgst": false 00:39:32.810 }, 00:39:32.810 "method": "bdev_nvme_attach_controller" 00:39:32.810 }' 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # asan_lib= 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # asan_lib= 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:39:32.810 09:15:21 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:32.810 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:39:32.811 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:39:32.811 fio-3.35 00:39:32.811 Starting 2 threads 00:39:42.806 00:39:42.806 filename0: (groupid=0, jobs=1): err= 0: pid=2195004: Wed Nov 6 09:15:32 2024 00:39:42.806 read: IOPS=96, BW=386KiB/s (395kB/s)(3872KiB/10036msec) 00:39:42.806 slat (nsec): min=5436, max=37835, avg=6282.20, stdev=1572.08 00:39:42.806 clat (usec): min=40900, max=42968, avg=41451.23, stdev=526.96 00:39:42.806 lat (usec): min=40905, max=42974, avg=41457.51, stdev=527.14 00:39:42.806 clat percentiles (usec): 00:39:42.806 | 1.00th=[41157], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:39:42.806 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41681], 00:39:42.806 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:39:42.806 | 99.00th=[42730], 99.50th=[42730], 99.90th=[42730], 99.95th=[42730], 00:39:42.806 | 99.99th=[42730] 00:39:42.806 bw ( KiB/s): min= 352, max= 416, per=50.21%, avg=385.60, stdev=12.61, samples=20 00:39:42.806 iops : min= 88, max= 104, avg=96.40, stdev= 3.15, samples=20 00:39:42.806 lat (msec) : 50=100.00% 00:39:42.806 cpu : usr=95.45%, sys=4.34%, ctx=14, majf=0, minf=128 00:39:42.806 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:39:42.806 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:42.806 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:42.806 issued rwts: total=968,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:39:42.806 latency : target=0, window=0, percentile=100.00%, depth=4 00:39:42.806 filename1: (groupid=0, jobs=1): err= 0: pid=2195005: Wed Nov 6 09:15:32 2024 00:39:42.806 read: IOPS=95, BW=381KiB/s (390kB/s)(3824KiB/10036msec) 00:39:42.806 slat (nsec): min=5438, max=27831, avg=6586.20, stdev=1565.73 00:39:42.806 clat (usec): min=40926, max=42180, avg=41970.07, stdev=114.65 00:39:42.806 lat (usec): min=40934, max=42208, avg=41976.66, stdev=114.51 00:39:42.806 clat percentiles (usec): 00:39:42.806 | 1.00th=[41157], 5.00th=[42206], 10.00th=[42206], 20.00th=[42206], 00:39:42.806 | 30.00th=[42206], 40.00th=[42206], 50.00th=[42206], 60.00th=[42206], 00:39:42.806 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:39:42.806 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:39:42.806 | 99.99th=[42206] 00:39:42.806 bw ( KiB/s): min= 352, max= 384, per=49.55%, avg=380.80, stdev= 9.85, samples=20 00:39:42.806 iops : min= 88, max= 96, avg=95.20, stdev= 2.46, samples=20 00:39:42.806 lat (msec) : 50=100.00% 00:39:42.806 cpu : usr=95.65%, sys=4.15%, ctx=14, majf=0, minf=138 00:39:42.806 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:39:42.806 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:42.806 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:42.806 issued rwts: total=956,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:39:42.806 latency : target=0, window=0, percentile=100.00%, depth=4 00:39:42.806 00:39:42.806 Run status group 0 (all jobs): 00:39:42.806 READ: bw=767KiB/s (785kB/s), 381KiB/s-386KiB/s (390kB/s-395kB/s), io=7696KiB (7881kB), run=10036-10036msec 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@96 -- # destroy_subsystems 0 1 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@43 -- # local sub 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@45 -- # for sub in "$@" 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@46 -- # destroy_subsystem 0 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@36 -- # local sub_id=0 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@45 -- # for sub in "$@" 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@46 -- # destroy_subsystem 1 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@36 -- # local sub_id=1 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:42.806 00:39:42.806 real 0m11.374s 00:39:42.806 user 0m36.471s 00:39:42.806 sys 0m1.269s 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1126 -- # xtrace_disable 00:39:42.806 09:15:32 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:39:42.806 ************************************ 00:39:42.806 END TEST fio_dif_1_multi_subsystems 00:39:42.806 ************************************ 00:39:43.067 09:15:32 nvmf_dif -- target/dif.sh@143 -- # run_test fio_dif_rand_params fio_dif_rand_params 00:39:43.067 09:15:32 nvmf_dif -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:39:43.067 09:15:32 nvmf_dif -- common/autotest_common.sh@1107 -- # xtrace_disable 00:39:43.067 09:15:32 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:39:43.067 ************************************ 00:39:43.068 START TEST fio_dif_rand_params 00:39:43.068 ************************************ 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1125 -- # fio_dif_rand_params 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@100 -- # local NULL_DIF 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@101 -- # local bs numjobs runtime iodepth files 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # NULL_DIF=3 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # bs=128k 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # numjobs=3 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # iodepth=3 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # runtime=5 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@105 -- # create_subsystems 0 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:43.068 bdev_null0 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:43.068 09:15:32 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:43.068 [2024-11-06 09:15:33.017203] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- target/dif.sh@106 -- # fio /dev/fd/62 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- target/dif.sh@106 -- # create_json_sub_conf 0 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # config=() 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # local subsystem config 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:39:43.068 { 00:39:43.068 "params": { 00:39:43.068 "name": "Nvme$subsystem", 00:39:43.068 "trtype": "$TEST_TRANSPORT", 00:39:43.068 "traddr": "$NVMF_FIRST_TARGET_IP", 00:39:43.068 "adrfam": "ipv4", 00:39:43.068 "trsvcid": "$NVMF_PORT", 00:39:43.068 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:39:43.068 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:39:43.068 "hdgst": ${hdgst:-false}, 00:39:43.068 "ddgst": ${ddgst:-false} 00:39:43.068 }, 00:39:43.068 "method": "bdev_nvme_attach_controller" 00:39:43.068 } 00:39:43.068 EOF 00:39:43.068 )") 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # local sanitizers 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # shift 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local asan_lib= 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libasan 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # jq . 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@583 -- # IFS=, 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:39:43.068 "params": { 00:39:43.068 "name": "Nvme0", 00:39:43.068 "trtype": "tcp", 00:39:43.068 "traddr": "10.0.0.2", 00:39:43.068 "adrfam": "ipv4", 00:39:43.068 "trsvcid": "4420", 00:39:43.068 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:39:43.068 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:39:43.068 "hdgst": false, 00:39:43.068 "ddgst": false 00:39:43.068 }, 00:39:43.068 "method": "bdev_nvme_attach_controller" 00:39:43.068 }' 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:39:43.068 09:15:33 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:43.638 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:39:43.638 ... 00:39:43.638 fio-3.35 00:39:43.638 Starting 3 threads 00:39:50.225 00:39:50.225 filename0: (groupid=0, jobs=1): err= 0: pid=2197378: Wed Nov 6 09:15:39 2024 00:39:50.225 read: IOPS=238, BW=29.8MiB/s (31.3MB/s)(149MiB/5005msec) 00:39:50.225 slat (nsec): min=5476, max=37328, avg=7390.09, stdev=2070.56 00:39:50.225 clat (usec): min=4999, max=53228, avg=12562.49, stdev=6744.01 00:39:50.225 lat (usec): min=5004, max=53235, avg=12569.88, stdev=6744.07 00:39:50.225 clat percentiles (usec): 00:39:50.225 | 1.00th=[ 6325], 5.00th=[ 7832], 10.00th=[ 8455], 20.00th=[ 9896], 00:39:50.225 | 30.00th=[10421], 40.00th=[11076], 50.00th=[11731], 60.00th=[12256], 00:39:50.225 | 70.00th=[12780], 80.00th=[13435], 90.00th=[14353], 95.00th=[15401], 00:39:50.225 | 99.00th=[51119], 99.50th=[52167], 99.90th=[52691], 99.95th=[53216], 00:39:50.225 | 99.99th=[53216] 00:39:50.225 bw ( KiB/s): min=27392, max=35584, per=34.43%, avg=30515.20, stdev=2890.77, samples=10 00:39:50.225 iops : min= 214, max= 278, avg=238.40, stdev=22.58, samples=10 00:39:50.225 lat (msec) : 10=22.53%, 20=74.71%, 50=0.84%, 100=1.93% 00:39:50.225 cpu : usr=94.54%, sys=5.18%, ctx=9, majf=0, minf=83 00:39:50.225 IO depths : 1=0.2%, 2=99.8%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:39:50.225 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:50.225 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:50.225 issued rwts: total=1194,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:39:50.225 latency : target=0, window=0, percentile=100.00%, depth=3 00:39:50.225 filename0: (groupid=0, jobs=1): err= 0: pid=2197379: Wed Nov 6 09:15:39 2024 00:39:50.225 read: IOPS=236, BW=29.6MiB/s (31.0MB/s)(150MiB/5052msec) 00:39:50.225 slat (nsec): min=5503, max=32386, avg=7399.39, stdev=1785.93 00:39:50.225 clat (usec): min=4809, max=55246, avg=12620.47, stdev=6195.66 00:39:50.225 lat (usec): min=4815, max=55255, avg=12627.87, stdev=6195.65 00:39:50.225 clat percentiles (usec): 00:39:50.225 | 1.00th=[ 6390], 5.00th=[ 7963], 10.00th=[ 8979], 20.00th=[10028], 00:39:50.225 | 30.00th=[10814], 40.00th=[11338], 50.00th=[11863], 60.00th=[12387], 00:39:50.225 | 70.00th=[12911], 80.00th=[13829], 90.00th=[14877], 95.00th=[15533], 00:39:50.225 | 99.00th=[51643], 99.50th=[53740], 99.90th=[54789], 99.95th=[55313], 00:39:50.225 | 99.99th=[55313] 00:39:50.225 bw ( KiB/s): min=26112, max=33024, per=34.50%, avg=30572.50, stdev=1791.47, samples=10 00:39:50.225 iops : min= 204, max= 258, avg=238.80, stdev=13.99, samples=10 00:39:50.225 lat (msec) : 10=19.73%, 20=78.09%, 50=0.50%, 100=1.67% 00:39:50.225 cpu : usr=94.04%, sys=5.70%, ctx=11, majf=0, minf=77 00:39:50.225 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:39:50.225 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:50.225 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:50.225 issued rwts: total=1196,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:39:50.225 latency : target=0, window=0, percentile=100.00%, depth=3 00:39:50.225 filename0: (groupid=0, jobs=1): err= 0: pid=2197380: Wed Nov 6 09:15:39 2024 00:39:50.225 read: IOPS=219, BW=27.4MiB/s (28.7MB/s)(139MiB/5052msec) 00:39:50.225 slat (nsec): min=5474, max=66046, avg=7626.49, stdev=2508.74 00:39:50.225 clat (usec): min=5973, max=56059, avg=13628.65, stdev=7567.54 00:39:50.225 lat (usec): min=5980, max=56064, avg=13636.28, stdev=7567.54 00:39:50.225 clat percentiles (usec): 00:39:50.225 | 1.00th=[ 6849], 5.00th=[ 8291], 10.00th=[ 9241], 20.00th=[10421], 00:39:50.225 | 30.00th=[11338], 40.00th=[11863], 50.00th=[12387], 60.00th=[12911], 00:39:50.225 | 70.00th=[13698], 80.00th=[14484], 90.00th=[15664], 95.00th=[16450], 00:39:50.225 | 99.00th=[52691], 99.50th=[54264], 99.90th=[55313], 99.95th=[55837], 00:39:50.225 | 99.99th=[55837] 00:39:50.225 bw ( KiB/s): min=21504, max=34560, per=31.92%, avg=28288.00, stdev=4161.70, samples=10 00:39:50.225 iops : min= 168, max= 270, avg=221.00, stdev=32.51, samples=10 00:39:50.225 lat (msec) : 10=16.06%, 20=80.42%, 50=0.90%, 100=2.62% 00:39:50.225 cpu : usr=94.02%, sys=5.72%, ctx=8, majf=0, minf=164 00:39:50.225 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:39:50.225 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:50.225 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:39:50.225 issued rwts: total=1108,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:39:50.225 latency : target=0, window=0, percentile=100.00%, depth=3 00:39:50.225 00:39:50.225 Run status group 0 (all jobs): 00:39:50.226 READ: bw=86.5MiB/s (90.8MB/s), 27.4MiB/s-29.8MiB/s (28.7MB/s-31.3MB/s), io=437MiB (458MB), run=5005-5052msec 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@107 -- # destroy_subsystems 0 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # NULL_DIF=2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # bs=4k 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # numjobs=8 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # iodepth=16 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # runtime= 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # files=2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@111 -- # create_subsystems 0 1 2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 bdev_null0 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 [2024-11-06 09:15:39.340249] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 1 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=1 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 bdev_null1 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null2 64 512 --md-size 16 --dif-type 2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 bdev_null2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 --serial-number 53313233-2 --allow-any-host 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 bdev_null2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@112 -- # fio /dev/fd/62 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@112 -- # create_json_sub_conf 0 1 2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 2 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # config=() 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # local subsystem config 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:39:50.226 { 00:39:50.226 "params": { 00:39:50.226 "name": "Nvme$subsystem", 00:39:50.226 "trtype": "$TEST_TRANSPORT", 00:39:50.226 "traddr": "$NVMF_FIRST_TARGET_IP", 00:39:50.226 "adrfam": "ipv4", 00:39:50.226 "trsvcid": "$NVMF_PORT", 00:39:50.226 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:39:50.226 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:39:50.226 "hdgst": ${hdgst:-false}, 00:39:50.226 "ddgst": ${ddgst:-false} 00:39:50.226 }, 00:39:50.226 "method": "bdev_nvme_attach_controller" 00:39:50.226 } 00:39:50.226 EOF 00:39:50.226 )") 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # local sanitizers 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # shift 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local asan_lib= 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libasan 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:39:50.226 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:39:50.226 { 00:39:50.226 "params": { 00:39:50.226 "name": "Nvme$subsystem", 00:39:50.226 "trtype": "$TEST_TRANSPORT", 00:39:50.227 "traddr": "$NVMF_FIRST_TARGET_IP", 00:39:50.227 "adrfam": "ipv4", 00:39:50.227 "trsvcid": "$NVMF_PORT", 00:39:50.227 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:39:50.227 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:39:50.227 "hdgst": ${hdgst:-false}, 00:39:50.227 "ddgst": ${ddgst:-false} 00:39:50.227 }, 00:39:50.227 "method": "bdev_nvme_attach_controller" 00:39:50.227 } 00:39:50.227 EOF 00:39:50.227 )") 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:39:50.227 { 00:39:50.227 "params": { 00:39:50.227 "name": "Nvme$subsystem", 00:39:50.227 "trtype": "$TEST_TRANSPORT", 00:39:50.227 "traddr": "$NVMF_FIRST_TARGET_IP", 00:39:50.227 "adrfam": "ipv4", 00:39:50.227 "trsvcid": "$NVMF_PORT", 00:39:50.227 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:39:50.227 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:39:50.227 "hdgst": ${hdgst:-false}, 00:39:50.227 "ddgst": ${ddgst:-false} 00:39:50.227 }, 00:39:50.227 "method": "bdev_nvme_attach_controller" 00:39:50.227 } 00:39:50.227 EOF 00:39:50.227 )") 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # jq . 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@583 -- # IFS=, 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:39:50.227 "params": { 00:39:50.227 "name": "Nvme0", 00:39:50.227 "trtype": "tcp", 00:39:50.227 "traddr": "10.0.0.2", 00:39:50.227 "adrfam": "ipv4", 00:39:50.227 "trsvcid": "4420", 00:39:50.227 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:39:50.227 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:39:50.227 "hdgst": false, 00:39:50.227 "ddgst": false 00:39:50.227 }, 00:39:50.227 "method": "bdev_nvme_attach_controller" 00:39:50.227 },{ 00:39:50.227 "params": { 00:39:50.227 "name": "Nvme1", 00:39:50.227 "trtype": "tcp", 00:39:50.227 "traddr": "10.0.0.2", 00:39:50.227 "adrfam": "ipv4", 00:39:50.227 "trsvcid": "4420", 00:39:50.227 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:39:50.227 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:39:50.227 "hdgst": false, 00:39:50.227 "ddgst": false 00:39:50.227 }, 00:39:50.227 "method": "bdev_nvme_attach_controller" 00:39:50.227 },{ 00:39:50.227 "params": { 00:39:50.227 "name": "Nvme2", 00:39:50.227 "trtype": "tcp", 00:39:50.227 "traddr": "10.0.0.2", 00:39:50.227 "adrfam": "ipv4", 00:39:50.227 "trsvcid": "4420", 00:39:50.227 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:39:50.227 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:39:50.227 "hdgst": false, 00:39:50.227 "ddgst": false 00:39:50.227 }, 00:39:50.227 "method": "bdev_nvme_attach_controller" 00:39:50.227 }' 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:39:50.227 09:15:39 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:39:50.227 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:39:50.227 ... 00:39:50.227 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:39:50.227 ... 00:39:50.227 filename2: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:39:50.227 ... 00:39:50.227 fio-3.35 00:39:50.227 Starting 24 threads 00:40:02.466 00:40:02.466 filename0: (groupid=0, jobs=1): err= 0: pid=2198782: Wed Nov 6 09:15:51 2024 00:40:02.466 read: IOPS=503, BW=2013KiB/s (2062kB/s)(19.7MiB/10001msec) 00:40:02.466 slat (usec): min=5, max=102, avg=13.35, stdev=10.93 00:40:02.466 clat (usec): min=8690, max=57665, avg=31681.20, stdev=4522.27 00:40:02.466 lat (usec): min=8696, max=57672, avg=31694.55, stdev=4522.93 00:40:02.466 clat percentiles (usec): 00:40:02.466 | 1.00th=[12518], 5.00th=[20055], 10.00th=[27657], 20.00th=[32113], 00:40:02.466 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.466 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.466 | 99.00th=[35914], 99.50th=[39060], 99.90th=[57410], 99.95th=[57410], 00:40:02.466 | 99.99th=[57410] 00:40:02.466 bw ( KiB/s): min= 1792, max= 2416, per=4.29%, avg=2009.26, stdev=140.43, samples=19 00:40:02.466 iops : min= 448, max= 604, avg=502.32, stdev=35.11, samples=19 00:40:02.466 lat (msec) : 10=0.60%, 20=4.37%, 50=94.87%, 100=0.16% 00:40:02.466 cpu : usr=98.85%, sys=0.82%, ctx=18, majf=0, minf=9 00:40:02.466 IO depths : 1=5.4%, 2=11.0%, 4=22.9%, 8=53.5%, 16=7.2%, 32=0.0%, >=64=0.0% 00:40:02.466 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.466 complete : 0=0.0%, 4=93.5%, 8=0.7%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.466 issued rwts: total=5034,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.466 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.466 filename0: (groupid=0, jobs=1): err= 0: pid=2198783: Wed Nov 6 09:15:51 2024 00:40:02.466 read: IOPS=496, BW=1985KiB/s (2032kB/s)(19.4MiB/10001msec) 00:40:02.466 slat (usec): min=5, max=121, avg=22.12, stdev=18.76 00:40:02.466 clat (usec): min=12609, max=55429, avg=32044.84, stdev=4071.45 00:40:02.466 lat (usec): min=12616, max=55484, avg=32066.96, stdev=4073.74 00:40:02.466 clat percentiles (usec): 00:40:02.466 | 1.00th=[16319], 5.00th=[22414], 10.00th=[30016], 20.00th=[32113], 00:40:02.466 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32637], 60.00th=[32900], 00:40:02.466 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.466 | 99.00th=[43254], 99.50th=[52167], 99.90th=[54264], 99.95th=[54264], 00:40:02.466 | 99.99th=[55313] 00:40:02.466 bw ( KiB/s): min= 1920, max= 2533, per=4.24%, avg=1988.47, stdev=148.83, samples=19 00:40:02.466 iops : min= 480, max= 633, avg=497.11, stdev=37.16, samples=19 00:40:02.466 lat (msec) : 20=2.74%, 50=96.63%, 100=0.62% 00:40:02.466 cpu : usr=98.94%, sys=0.74%, ctx=13, majf=0, minf=9 00:40:02.466 IO depths : 1=4.8%, 2=9.9%, 4=21.8%, 8=55.7%, 16=7.8%, 32=0.0%, >=64=0.0% 00:40:02.466 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.466 complete : 0=0.0%, 4=93.3%, 8=1.0%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.466 issued rwts: total=4962,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.466 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.466 filename0: (groupid=0, jobs=1): err= 0: pid=2198784: Wed Nov 6 09:15:51 2024 00:40:02.466 read: IOPS=495, BW=1981KiB/s (2029kB/s)(19.4MiB/10006msec) 00:40:02.466 slat (usec): min=5, max=116, avg=20.32, stdev=18.09 00:40:02.466 clat (usec): min=6521, max=73418, avg=32188.52, stdev=4913.02 00:40:02.466 lat (usec): min=6527, max=73434, avg=32208.83, stdev=4914.99 00:40:02.466 clat percentiles (usec): 00:40:02.466 | 1.00th=[19006], 5.00th=[22152], 10.00th=[25297], 20.00th=[31851], 00:40:02.466 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32637], 60.00th=[32900], 00:40:02.466 | 70.00th=[33162], 80.00th=[33817], 90.00th=[34341], 95.00th=[38536], 00:40:02.466 | 99.00th=[49021], 99.50th=[52167], 99.90th=[55837], 99.95th=[55837], 00:40:02.466 | 99.99th=[73925] 00:40:02.466 bw ( KiB/s): min= 1715, max= 2160, per=4.22%, avg=1976.15, stdev=89.83, samples=20 00:40:02.466 iops : min= 428, max= 540, avg=494.00, stdev=22.57, samples=20 00:40:02.466 lat (msec) : 10=0.18%, 20=1.21%, 50=97.70%, 100=0.91% 00:40:02.466 cpu : usr=98.85%, sys=0.82%, ctx=24, majf=0, minf=9 00:40:02.466 IO depths : 1=0.6%, 2=2.2%, 4=8.1%, 8=74.2%, 16=14.9%, 32=0.0%, >=64=0.0% 00:40:02.466 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.466 complete : 0=0.0%, 4=90.5%, 8=6.7%, 16=2.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.466 issued rwts: total=4956,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.466 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.466 filename0: (groupid=0, jobs=1): err= 0: pid=2198785: Wed Nov 6 09:15:51 2024 00:40:02.466 read: IOPS=482, BW=1931KiB/s (1978kB/s)(18.9MiB/10007msec) 00:40:02.466 slat (usec): min=5, max=117, avg=27.14, stdev=19.31 00:40:02.466 clat (usec): min=19083, max=41707, avg=32878.94, stdev=1144.88 00:40:02.466 lat (usec): min=19096, max=41745, avg=32906.09, stdev=1144.64 00:40:02.466 clat percentiles (usec): 00:40:02.466 | 1.00th=[31065], 5.00th=[31851], 10.00th=[32113], 20.00th=[32375], 00:40:02.466 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32900], 60.00th=[32900], 00:40:02.466 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.466 | 99.00th=[35914], 99.50th=[35914], 99.90th=[38011], 99.95th=[38011], 00:40:02.466 | 99.99th=[41681] 00:40:02.467 bw ( KiB/s): min= 1920, max= 2048, per=4.11%, avg=1926.74, stdev=29.37, samples=19 00:40:02.467 iops : min= 480, max= 512, avg=481.68, stdev= 7.34, samples=19 00:40:02.467 lat (msec) : 20=0.33%, 50=99.67% 00:40:02.467 cpu : usr=98.95%, sys=0.71%, ctx=21, majf=0, minf=9 00:40:02.467 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.3%, 32=0.0%, >=64=0.0% 00:40:02.467 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 issued rwts: total=4832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.467 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.467 filename0: (groupid=0, jobs=1): err= 0: pid=2198787: Wed Nov 6 09:15:51 2024 00:40:02.467 read: IOPS=485, BW=1942KiB/s (1989kB/s)(19.0MiB/10016msec) 00:40:02.467 slat (usec): min=5, max=125, avg=18.40, stdev=13.68 00:40:02.467 clat (usec): min=14778, max=53353, avg=32798.14, stdev=2042.13 00:40:02.467 lat (usec): min=14786, max=53386, avg=32816.54, stdev=2042.34 00:40:02.467 clat percentiles (usec): 00:40:02.467 | 1.00th=[21890], 5.00th=[31851], 10.00th=[32113], 20.00th=[32375], 00:40:02.467 | 30.00th=[32637], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.467 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.467 | 99.00th=[35390], 99.50th=[36439], 99.90th=[43779], 99.95th=[43779], 00:40:02.467 | 99.99th=[53216] 00:40:02.467 bw ( KiB/s): min= 1920, max= 2052, per=4.14%, avg=1939.40, stdev=47.39, samples=20 00:40:02.467 iops : min= 480, max= 513, avg=484.85, stdev=11.85, samples=20 00:40:02.467 lat (msec) : 20=0.86%, 50=99.10%, 100=0.04% 00:40:02.467 cpu : usr=98.70%, sys=0.97%, ctx=15, majf=0, minf=9 00:40:02.467 IO depths : 1=6.0%, 2=12.2%, 4=24.7%, 8=50.6%, 16=6.5%, 32=0.0%, >=64=0.0% 00:40:02.467 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 complete : 0=0.0%, 4=94.0%, 8=0.1%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 issued rwts: total=4864,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.467 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.467 filename0: (groupid=0, jobs=1): err= 0: pid=2198788: Wed Nov 6 09:15:51 2024 00:40:02.467 read: IOPS=486, BW=1948KiB/s (1994kB/s)(19.1MiB/10022msec) 00:40:02.467 slat (usec): min=5, max=119, avg=19.40, stdev=17.90 00:40:02.467 clat (usec): min=13164, max=36427, avg=32702.07, stdev=2257.42 00:40:02.467 lat (usec): min=13183, max=36437, avg=32721.48, stdev=2256.28 00:40:02.467 clat percentiles (usec): 00:40:02.467 | 1.00th=[18482], 5.00th=[31851], 10.00th=[32113], 20.00th=[32375], 00:40:02.467 | 30.00th=[32637], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.467 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.467 | 99.00th=[35390], 99.50th=[35914], 99.90th=[36439], 99.95th=[36439], 00:40:02.467 | 99.99th=[36439] 00:40:02.467 bw ( KiB/s): min= 1920, max= 2176, per=4.15%, avg=1945.60, stdev=66.96, samples=20 00:40:02.467 iops : min= 480, max= 544, avg=486.40, stdev=16.74, samples=20 00:40:02.467 lat (msec) : 20=1.31%, 50=98.69% 00:40:02.467 cpu : usr=98.94%, sys=0.65%, ctx=78, majf=0, minf=9 00:40:02.467 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:40:02.467 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 issued rwts: total=4880,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.467 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.467 filename0: (groupid=0, jobs=1): err= 0: pid=2198789: Wed Nov 6 09:15:51 2024 00:40:02.467 read: IOPS=492, BW=1972KiB/s (2019kB/s)(19.3MiB/10021msec) 00:40:02.467 slat (usec): min=5, max=122, avg=19.21, stdev=16.04 00:40:02.467 clat (usec): min=16135, max=52879, avg=32293.82, stdev=4805.30 00:40:02.467 lat (usec): min=16141, max=52886, avg=32313.03, stdev=4808.21 00:40:02.467 clat percentiles (usec): 00:40:02.467 | 1.00th=[19792], 5.00th=[22414], 10.00th=[25822], 20.00th=[31851], 00:40:02.467 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32637], 60.00th=[32900], 00:40:02.467 | 70.00th=[33162], 80.00th=[33424], 90.00th=[34341], 95.00th=[40109], 00:40:02.467 | 99.00th=[50594], 99.50th=[52691], 99.90th=[52691], 99.95th=[52691], 00:40:02.467 | 99.99th=[52691] 00:40:02.467 bw ( KiB/s): min= 1792, max= 2144, per=4.21%, avg=1972.00, stdev=95.84, samples=20 00:40:02.467 iops : min= 448, max= 536, avg=493.00, stdev=23.96, samples=20 00:40:02.467 lat (msec) : 20=1.23%, 50=97.69%, 100=1.07% 00:40:02.467 cpu : usr=98.88%, sys=0.79%, ctx=16, majf=0, minf=9 00:40:02.467 IO depths : 1=1.9%, 2=6.2%, 4=19.3%, 8=61.8%, 16=10.8%, 32=0.0%, >=64=0.0% 00:40:02.467 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 complete : 0=0.0%, 4=92.8%, 8=1.7%, 16=5.5%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 issued rwts: total=4940,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.467 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.467 filename0: (groupid=0, jobs=1): err= 0: pid=2198790: Wed Nov 6 09:15:51 2024 00:40:02.467 read: IOPS=483, BW=1933KiB/s (1979kB/s)(18.9MiB/10003msec) 00:40:02.467 slat (usec): min=5, max=108, avg=23.22, stdev=15.17 00:40:02.467 clat (usec): min=12472, max=65197, avg=32912.70, stdev=3107.08 00:40:02.467 lat (usec): min=12478, max=65213, avg=32935.91, stdev=3107.24 00:40:02.467 clat percentiles (usec): 00:40:02.467 | 1.00th=[21627], 5.00th=[31851], 10.00th=[32113], 20.00th=[32375], 00:40:02.467 | 30.00th=[32637], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.467 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.467 | 99.00th=[44303], 99.50th=[50070], 99.90th=[65274], 99.95th=[65274], 00:40:02.467 | 99.99th=[65274] 00:40:02.467 bw ( KiB/s): min= 1792, max= 2048, per=4.10%, avg=1920.84, stdev=67.14, samples=19 00:40:02.467 iops : min= 448, max= 512, avg=480.21, stdev=16.78, samples=19 00:40:02.467 lat (msec) : 20=0.79%, 50=98.68%, 100=0.54% 00:40:02.467 cpu : usr=98.83%, sys=0.83%, ctx=61, majf=0, minf=9 00:40:02.467 IO depths : 1=5.9%, 2=11.9%, 4=24.2%, 8=51.3%, 16=6.7%, 32=0.0%, >=64=0.0% 00:40:02.467 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 complete : 0=0.0%, 4=93.9%, 8=0.3%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 issued rwts: total=4834,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.467 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.467 filename1: (groupid=0, jobs=1): err= 0: pid=2198791: Wed Nov 6 09:15:51 2024 00:40:02.467 read: IOPS=489, BW=1957KiB/s (2004kB/s)(19.1MiB/10013msec) 00:40:02.467 slat (usec): min=5, max=125, avg=25.94, stdev=19.96 00:40:02.467 clat (usec): min=16793, max=51968, avg=32476.18, stdev=3353.38 00:40:02.467 lat (usec): min=16800, max=51983, avg=32502.12, stdev=3355.97 00:40:02.467 clat percentiles (usec): 00:40:02.467 | 1.00th=[20579], 5.00th=[25035], 10.00th=[31589], 20.00th=[32113], 00:40:02.467 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32637], 60.00th=[32900], 00:40:02.467 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34866], 00:40:02.467 | 99.00th=[43779], 99.50th=[47973], 99.90th=[51643], 99.95th=[52167], 00:40:02.467 | 99.99th=[52167] 00:40:02.467 bw ( KiB/s): min= 1920, max= 2096, per=4.17%, avg=1954.53, stdev=62.03, samples=19 00:40:02.467 iops : min= 480, max= 524, avg=488.63, stdev=15.51, samples=19 00:40:02.467 lat (msec) : 20=0.41%, 50=99.31%, 100=0.29% 00:40:02.467 cpu : usr=99.03%, sys=0.63%, ctx=15, majf=0, minf=9 00:40:02.467 IO depths : 1=4.2%, 2=9.4%, 4=21.7%, 8=56.3%, 16=8.4%, 32=0.0%, >=64=0.0% 00:40:02.467 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 complete : 0=0.0%, 4=93.4%, 8=1.0%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 issued rwts: total=4898,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.467 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.467 filename1: (groupid=0, jobs=1): err= 0: pid=2198792: Wed Nov 6 09:15:51 2024 00:40:02.467 read: IOPS=487, BW=1951KiB/s (1998kB/s)(19.1MiB/10017msec) 00:40:02.467 slat (nsec): min=5460, max=80972, avg=13235.79, stdev=10081.15 00:40:02.467 clat (usec): min=10890, max=45092, avg=32694.64, stdev=2278.79 00:40:02.467 lat (usec): min=10898, max=45098, avg=32707.88, stdev=2279.26 00:40:02.467 clat percentiles (usec): 00:40:02.467 | 1.00th=[21627], 5.00th=[31589], 10.00th=[32113], 20.00th=[32375], 00:40:02.467 | 30.00th=[32637], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.467 | 70.00th=[33424], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.467 | 99.00th=[35390], 99.50th=[35914], 99.90th=[42730], 99.95th=[42730], 00:40:02.467 | 99.99th=[45351] 00:40:02.467 bw ( KiB/s): min= 1920, max= 2144, per=4.16%, avg=1948.00, stdev=62.27, samples=20 00:40:02.467 iops : min= 480, max= 536, avg=487.00, stdev=15.57, samples=20 00:40:02.467 lat (msec) : 20=0.86%, 50=99.14% 00:40:02.467 cpu : usr=98.78%, sys=0.88%, ctx=13, majf=0, minf=9 00:40:02.467 IO depths : 1=6.0%, 2=12.1%, 4=24.4%, 8=50.9%, 16=6.5%, 32=0.0%, >=64=0.0% 00:40:02.467 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 complete : 0=0.0%, 4=94.0%, 8=0.2%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 issued rwts: total=4886,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.467 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.467 filename1: (groupid=0, jobs=1): err= 0: pid=2198794: Wed Nov 6 09:15:51 2024 00:40:02.467 read: IOPS=486, BW=1947KiB/s (1994kB/s)(19.0MiB/10006msec) 00:40:02.467 slat (usec): min=5, max=107, avg=24.32, stdev=17.93 00:40:02.467 clat (usec): min=6427, max=78709, avg=32645.21, stdev=4007.60 00:40:02.467 lat (usec): min=6438, max=78728, avg=32669.52, stdev=4008.43 00:40:02.467 clat percentiles (usec): 00:40:02.467 | 1.00th=[20841], 5.00th=[26608], 10.00th=[31589], 20.00th=[32375], 00:40:02.467 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32637], 60.00th=[32900], 00:40:02.467 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[35390], 00:40:02.467 | 99.00th=[40633], 99.50th=[42730], 99.90th=[79168], 99.95th=[79168], 00:40:02.467 | 99.99th=[79168] 00:40:02.467 bw ( KiB/s): min= 1664, max= 2064, per=4.13%, avg=1936.00, stdev=81.58, samples=19 00:40:02.467 iops : min= 416, max= 516, avg=484.00, stdev=20.40, samples=19 00:40:02.467 lat (msec) : 10=0.12%, 20=0.53%, 50=99.01%, 100=0.33% 00:40:02.467 cpu : usr=98.87%, sys=0.78%, ctx=25, majf=0, minf=9 00:40:02.467 IO depths : 1=4.4%, 2=8.9%, 4=18.5%, 8=59.1%, 16=9.2%, 32=0.0%, >=64=0.0% 00:40:02.467 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 complete : 0=0.0%, 4=92.5%, 8=2.7%, 16=4.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.467 issued rwts: total=4870,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.467 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.467 filename1: (groupid=0, jobs=1): err= 0: pid=2198795: Wed Nov 6 09:15:51 2024 00:40:02.468 read: IOPS=488, BW=1954KiB/s (2001kB/s)(19.1MiB/10006msec) 00:40:02.468 slat (usec): min=5, max=114, avg=24.07, stdev=17.09 00:40:02.468 clat (usec): min=8717, max=56830, avg=32531.92, stdev=4077.51 00:40:02.468 lat (usec): min=8725, max=56855, avg=32556.00, stdev=4079.48 00:40:02.468 clat percentiles (usec): 00:40:02.468 | 1.00th=[19006], 5.00th=[25297], 10.00th=[30540], 20.00th=[32113], 00:40:02.468 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32637], 60.00th=[32900], 00:40:02.468 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34866], 00:40:02.468 | 99.00th=[50070], 99.50th=[53740], 99.90th=[56886], 99.95th=[56886], 00:40:02.468 | 99.99th=[56886] 00:40:02.468 bw ( KiB/s): min= 1715, max= 2080, per=4.16%, avg=1948.95, stdev=91.38, samples=20 00:40:02.468 iops : min= 428, max= 520, avg=487.20, stdev=22.95, samples=20 00:40:02.468 lat (msec) : 10=0.12%, 20=1.21%, 50=97.65%, 100=1.02% 00:40:02.468 cpu : usr=98.96%, sys=0.69%, ctx=14, majf=0, minf=9 00:40:02.468 IO depths : 1=4.7%, 2=9.7%, 4=21.3%, 8=56.3%, 16=8.0%, 32=0.0%, >=64=0.0% 00:40:02.468 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 complete : 0=0.0%, 4=93.2%, 8=1.2%, 16=5.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 issued rwts: total=4888,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.468 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.468 filename1: (groupid=0, jobs=1): err= 0: pid=2198796: Wed Nov 6 09:15:51 2024 00:40:02.468 read: IOPS=506, BW=2026KiB/s (2075kB/s)(19.8MiB/10022msec) 00:40:02.468 slat (usec): min=5, max=124, avg=13.70, stdev=11.73 00:40:02.468 clat (usec): min=10715, max=57200, avg=31482.91, stdev=4805.67 00:40:02.468 lat (usec): min=10721, max=57210, avg=31496.61, stdev=4807.30 00:40:02.468 clat percentiles (usec): 00:40:02.468 | 1.00th=[15139], 5.00th=[20055], 10.00th=[24249], 20.00th=[32113], 00:40:02.468 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32637], 60.00th=[32900], 00:40:02.468 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.468 | 99.00th=[41681], 99.50th=[52167], 99.90th=[56361], 99.95th=[57410], 00:40:02.468 | 99.99th=[57410] 00:40:02.468 bw ( KiB/s): min= 1792, max= 2480, per=4.32%, avg=2024.00, stdev=165.26, samples=20 00:40:02.468 iops : min= 448, max= 620, avg=506.00, stdev=41.32, samples=20 00:40:02.468 lat (msec) : 20=4.94%, 50=94.42%, 100=0.63% 00:40:02.468 cpu : usr=98.77%, sys=0.89%, ctx=12, majf=0, minf=9 00:40:02.468 IO depths : 1=4.7%, 2=9.7%, 4=21.2%, 8=56.5%, 16=7.9%, 32=0.0%, >=64=0.0% 00:40:02.468 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 complete : 0=0.0%, 4=93.1%, 8=1.2%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 issued rwts: total=5076,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.468 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.468 filename1: (groupid=0, jobs=1): err= 0: pid=2198797: Wed Nov 6 09:15:51 2024 00:40:02.468 read: IOPS=482, BW=1931KiB/s (1977kB/s)(18.9MiB/10011msec) 00:40:02.468 slat (usec): min=5, max=111, avg=20.56, stdev=14.56 00:40:02.468 clat (usec): min=12237, max=59862, avg=32979.27, stdev=2232.00 00:40:02.468 lat (usec): min=12246, max=59877, avg=32999.83, stdev=2230.88 00:40:02.468 clat percentiles (usec): 00:40:02.468 | 1.00th=[30278], 5.00th=[32113], 10.00th=[32113], 20.00th=[32375], 00:40:02.468 | 30.00th=[32637], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.468 | 70.00th=[33424], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.468 | 99.00th=[35914], 99.50th=[35914], 99.90th=[60031], 99.95th=[60031], 00:40:02.468 | 99.99th=[60031] 00:40:02.468 bw ( KiB/s): min= 1792, max= 2048, per=4.11%, avg=1925.35, stdev=47.93, samples=20 00:40:02.468 iops : min= 448, max= 512, avg=481.30, stdev=11.90, samples=20 00:40:02.468 lat (msec) : 20=0.66%, 50=99.01%, 100=0.33% 00:40:02.468 cpu : usr=98.85%, sys=0.81%, ctx=13, majf=0, minf=9 00:40:02.468 IO depths : 1=6.2%, 2=12.4%, 4=25.0%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:40:02.468 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 issued rwts: total=4832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.468 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.468 filename1: (groupid=0, jobs=1): err= 0: pid=2198798: Wed Nov 6 09:15:51 2024 00:40:02.468 read: IOPS=485, BW=1942KiB/s (1989kB/s)(19.0MiB/10017msec) 00:40:02.468 slat (usec): min=5, max=114, avg=22.72, stdev=15.94 00:40:02.468 clat (usec): min=13063, max=46907, avg=32759.82, stdev=1870.48 00:40:02.468 lat (usec): min=13072, max=46930, avg=32782.53, stdev=1870.35 00:40:02.468 clat percentiles (usec): 00:40:02.468 | 1.00th=[23462], 5.00th=[31851], 10.00th=[32113], 20.00th=[32375], 00:40:02.468 | 30.00th=[32637], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.468 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.468 | 99.00th=[35390], 99.50th=[35914], 99.90th=[42730], 99.95th=[42730], 00:40:02.468 | 99.99th=[46924] 00:40:02.468 bw ( KiB/s): min= 1904, max= 2048, per=4.14%, avg=1939.20, stdev=47.18, samples=20 00:40:02.468 iops : min= 476, max= 512, avg=484.80, stdev=11.79, samples=20 00:40:02.468 lat (msec) : 20=0.66%, 50=99.34% 00:40:02.468 cpu : usr=98.74%, sys=0.92%, ctx=16, majf=0, minf=9 00:40:02.468 IO depths : 1=6.1%, 2=12.3%, 4=24.8%, 8=50.4%, 16=6.4%, 32=0.0%, >=64=0.0% 00:40:02.468 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 complete : 0=0.0%, 4=94.1%, 8=0.1%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 issued rwts: total=4864,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.468 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.468 filename1: (groupid=0, jobs=1): err= 0: pid=2198799: Wed Nov 6 09:15:51 2024 00:40:02.468 read: IOPS=488, BW=1956KiB/s (2003kB/s)(19.1MiB/10006msec) 00:40:02.468 slat (nsec): min=5464, max=83732, avg=20333.75, stdev=12064.94 00:40:02.468 clat (usec): min=12302, max=78920, avg=32551.25, stdev=4467.96 00:40:02.468 lat (usec): min=12308, max=78936, avg=32571.58, stdev=4469.17 00:40:02.468 clat percentiles (usec): 00:40:02.468 | 1.00th=[17957], 5.00th=[24511], 10.00th=[31327], 20.00th=[32375], 00:40:02.468 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.468 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[35390], 00:40:02.468 | 99.00th=[44303], 99.50th=[50594], 99.90th=[79168], 99.95th=[79168], 00:40:02.468 | 99.99th=[79168] 00:40:02.468 bw ( KiB/s): min= 1667, max= 2288, per=4.16%, avg=1950.55, stdev=111.65, samples=20 00:40:02.468 iops : min= 416, max= 572, avg=487.60, stdev=28.01, samples=20 00:40:02.468 lat (msec) : 20=1.80%, 50=97.63%, 100=0.57% 00:40:02.468 cpu : usr=98.88%, sys=0.79%, ctx=15, majf=0, minf=9 00:40:02.468 IO depths : 1=4.9%, 2=9.8%, 4=20.4%, 8=56.9%, 16=8.1%, 32=0.0%, >=64=0.0% 00:40:02.468 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 complete : 0=0.0%, 4=92.9%, 8=1.8%, 16=5.3%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 issued rwts: total=4892,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.468 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.468 filename2: (groupid=0, jobs=1): err= 0: pid=2198800: Wed Nov 6 09:15:51 2024 00:40:02.468 read: IOPS=486, BW=1944KiB/s (1991kB/s)(19.0MiB/10015msec) 00:40:02.468 slat (nsec): min=5456, max=90249, avg=18100.07, stdev=12831.33 00:40:02.468 clat (usec): min=14197, max=56402, avg=32750.91, stdev=3892.75 00:40:02.468 lat (usec): min=14204, max=56408, avg=32769.01, stdev=3894.11 00:40:02.468 clat percentiles (usec): 00:40:02.468 | 1.00th=[18482], 5.00th=[25560], 10.00th=[31851], 20.00th=[32375], 00:40:02.468 | 30.00th=[32637], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.468 | 70.00th=[33162], 80.00th=[33817], 90.00th=[33817], 95.00th=[35390], 00:40:02.468 | 99.00th=[49021], 99.50th=[52691], 99.90th=[56361], 99.95th=[56361], 00:40:02.468 | 99.99th=[56361] 00:40:02.468 bw ( KiB/s): min= 1920, max= 2048, per=4.13%, avg=1936.84, stdev=37.13, samples=19 00:40:02.468 iops : min= 480, max= 512, avg=484.21, stdev= 9.28, samples=19 00:40:02.468 lat (msec) : 20=1.83%, 50=97.35%, 100=0.82% 00:40:02.468 cpu : usr=98.81%, sys=0.87%, ctx=15, majf=0, minf=9 00:40:02.468 IO depths : 1=4.7%, 2=9.7%, 4=21.5%, 8=56.0%, 16=8.0%, 32=0.0%, >=64=0.0% 00:40:02.468 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 complete : 0=0.0%, 4=93.3%, 8=1.1%, 16=5.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 issued rwts: total=4868,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.468 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.468 filename2: (groupid=0, jobs=1): err= 0: pid=2198801: Wed Nov 6 09:15:51 2024 00:40:02.468 read: IOPS=497, BW=1988KiB/s (2036kB/s)(19.5MiB/10022msec) 00:40:02.468 slat (usec): min=5, max=122, avg=19.77, stdev=17.39 00:40:02.468 clat (usec): min=12481, max=60549, avg=32009.89, stdev=4655.29 00:40:02.468 lat (usec): min=12504, max=60607, avg=32029.66, stdev=4657.30 00:40:02.468 clat percentiles (usec): 00:40:02.468 | 1.00th=[15926], 5.00th=[21627], 10.00th=[26608], 20.00th=[32113], 00:40:02.468 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32637], 60.00th=[32900], 00:40:02.468 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.468 | 99.00th=[52167], 99.50th=[55837], 99.90th=[60556], 99.95th=[60556], 00:40:02.468 | 99.99th=[60556] 00:40:02.468 bw ( KiB/s): min= 1872, max= 2224, per=4.24%, avg=1986.40, stdev=96.74, samples=20 00:40:02.468 iops : min= 468, max= 556, avg=496.60, stdev=24.18, samples=20 00:40:02.468 lat (msec) : 20=3.19%, 50=95.72%, 100=1.08% 00:40:02.468 cpu : usr=98.78%, sys=0.86%, ctx=26, majf=0, minf=9 00:40:02.468 IO depths : 1=5.3%, 2=10.6%, 4=22.3%, 8=54.5%, 16=7.3%, 32=0.0%, >=64=0.0% 00:40:02.468 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 complete : 0=0.0%, 4=93.4%, 8=0.9%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.468 issued rwts: total=4982,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.468 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.468 filename2: (groupid=0, jobs=1): err= 0: pid=2198803: Wed Nov 6 09:15:51 2024 00:40:02.468 read: IOPS=489, BW=1958KiB/s (2005kB/s)(19.1MiB/10005msec) 00:40:02.468 slat (usec): min=5, max=126, avg=24.16, stdev=22.09 00:40:02.468 clat (usec): min=12415, max=65209, avg=32528.80, stdev=5682.91 00:40:02.468 lat (usec): min=12423, max=65226, avg=32552.96, stdev=5683.72 00:40:02.468 clat percentiles (usec): 00:40:02.468 | 1.00th=[17433], 5.00th=[21627], 10.00th=[25822], 20.00th=[31851], 00:40:02.468 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.468 | 70.00th=[33424], 80.00th=[33817], 90.00th=[35914], 95.00th=[42206], 00:40:02.468 | 99.00th=[52691], 99.50th=[56361], 99.90th=[65274], 99.95th=[65274], 00:40:02.468 | 99.99th=[65274] 00:40:02.468 bw ( KiB/s): min= 1712, max= 2112, per=4.16%, avg=1950.32, stdev=99.48, samples=19 00:40:02.469 iops : min= 428, max= 528, avg=487.58, stdev=24.87, samples=19 00:40:02.469 lat (msec) : 20=2.47%, 50=95.81%, 100=1.71% 00:40:02.469 cpu : usr=99.08%, sys=0.58%, ctx=34, majf=0, minf=9 00:40:02.469 IO depths : 1=0.4%, 2=2.1%, 4=10.0%, 8=72.8%, 16=14.6%, 32=0.0%, >=64=0.0% 00:40:02.469 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 complete : 0=0.0%, 4=91.0%, 8=5.8%, 16=3.2%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 issued rwts: total=4898,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.469 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.469 filename2: (groupid=0, jobs=1): err= 0: pid=2198804: Wed Nov 6 09:15:51 2024 00:40:02.469 read: IOPS=486, BW=1945KiB/s (1992kB/s)(19.0MiB/10014msec) 00:40:02.469 slat (usec): min=5, max=113, avg=26.99, stdev=17.08 00:40:02.469 clat (usec): min=14907, max=50519, avg=32653.10, stdev=2803.72 00:40:02.469 lat (usec): min=14922, max=50530, avg=32680.09, stdev=2804.31 00:40:02.469 clat percentiles (usec): 00:40:02.469 | 1.00th=[20841], 5.00th=[29754], 10.00th=[32113], 20.00th=[32375], 00:40:02.469 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32637], 60.00th=[32900], 00:40:02.469 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.469 | 99.00th=[43254], 99.50th=[44303], 99.90th=[50594], 99.95th=[50594], 00:40:02.469 | 99.99th=[50594] 00:40:02.469 bw ( KiB/s): min= 1920, max= 2048, per=4.14%, avg=1941.60, stdev=42.58, samples=20 00:40:02.469 iops : min= 480, max= 512, avg=485.40, stdev=10.64, samples=20 00:40:02.469 lat (msec) : 20=0.45%, 50=99.43%, 100=0.12% 00:40:02.469 cpu : usr=99.20%, sys=0.51%, ctx=14, majf=0, minf=9 00:40:02.469 IO depths : 1=4.7%, 2=10.6%, 4=23.8%, 8=53.2%, 16=7.8%, 32=0.0%, >=64=0.0% 00:40:02.469 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 complete : 0=0.0%, 4=93.9%, 8=0.3%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 issued rwts: total=4870,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.469 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.469 filename2: (groupid=0, jobs=1): err= 0: pid=2198805: Wed Nov 6 09:15:51 2024 00:40:02.469 read: IOPS=488, BW=1952KiB/s (1999kB/s)(19.1MiB/10012msec) 00:40:02.469 slat (usec): min=5, max=124, avg=26.93, stdev=19.37 00:40:02.469 clat (usec): min=12855, max=54351, avg=32557.59, stdev=3412.10 00:40:02.469 lat (usec): min=12864, max=54357, avg=32584.52, stdev=3413.81 00:40:02.469 clat percentiles (usec): 00:40:02.469 | 1.00th=[21103], 5.00th=[25035], 10.00th=[31851], 20.00th=[32113], 00:40:02.469 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32900], 60.00th=[32900], 00:40:02.469 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[35390], 00:40:02.469 | 99.00th=[43779], 99.50th=[44303], 99.90th=[54264], 99.95th=[54264], 00:40:02.469 | 99.99th=[54264] 00:40:02.469 bw ( KiB/s): min= 1888, max= 2176, per=4.16%, avg=1948.20, stdev=68.57, samples=20 00:40:02.469 iops : min= 472, max= 544, avg=487.05, stdev=17.14, samples=20 00:40:02.469 lat (msec) : 20=0.49%, 50=99.26%, 100=0.25% 00:40:02.469 cpu : usr=99.12%, sys=0.56%, ctx=15, majf=0, minf=9 00:40:02.469 IO depths : 1=4.2%, 2=9.1%, 4=21.4%, 8=56.8%, 16=8.5%, 32=0.0%, >=64=0.0% 00:40:02.469 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 complete : 0=0.0%, 4=93.3%, 8=1.2%, 16=5.5%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 issued rwts: total=4886,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.469 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.469 filename2: (groupid=0, jobs=1): err= 0: pid=2198806: Wed Nov 6 09:15:51 2024 00:40:02.469 read: IOPS=486, BW=1948KiB/s (1994kB/s)(19.1MiB/10022msec) 00:40:02.469 slat (usec): min=5, max=113, avg=23.52, stdev=19.80 00:40:02.469 clat (usec): min=12272, max=56484, avg=32638.49, stdev=2733.92 00:40:02.469 lat (usec): min=12279, max=56496, avg=32662.01, stdev=2734.92 00:40:02.469 clat percentiles (usec): 00:40:02.469 | 1.00th=[16188], 5.00th=[31851], 10.00th=[32113], 20.00th=[32375], 00:40:02.469 | 30.00th=[32637], 40.00th=[32637], 50.00th=[32900], 60.00th=[32900], 00:40:02.469 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.469 | 99.00th=[40109], 99.50th=[42206], 99.90th=[48497], 99.95th=[50594], 00:40:02.469 | 99.99th=[56361] 00:40:02.469 bw ( KiB/s): min= 1920, max= 2176, per=4.15%, avg=1945.60, stdev=65.54, samples=20 00:40:02.469 iops : min= 480, max= 544, avg=486.40, stdev=16.38, samples=20 00:40:02.469 lat (msec) : 20=1.43%, 50=98.48%, 100=0.08% 00:40:02.469 cpu : usr=99.03%, sys=0.63%, ctx=26, majf=0, minf=9 00:40:02.469 IO depths : 1=5.5%, 2=11.6%, 4=24.6%, 8=51.3%, 16=7.0%, 32=0.0%, >=64=0.0% 00:40:02.469 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 complete : 0=0.0%, 4=94.1%, 8=0.1%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 issued rwts: total=4880,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.469 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.469 filename2: (groupid=0, jobs=1): err= 0: pid=2198807: Wed Nov 6 09:15:51 2024 00:40:02.469 read: IOPS=482, BW=1932KiB/s (1978kB/s)(18.9MiB/10006msec) 00:40:02.469 slat (usec): min=5, max=110, avg=25.91, stdev=15.55 00:40:02.469 clat (usec): min=13877, max=54180, avg=32906.96, stdev=1953.02 00:40:02.469 lat (usec): min=13882, max=54196, avg=32932.87, stdev=1952.24 00:40:02.469 clat percentiles (usec): 00:40:02.469 | 1.00th=[30802], 5.00th=[31851], 10.00th=[32113], 20.00th=[32375], 00:40:02.469 | 30.00th=[32375], 40.00th=[32637], 50.00th=[32900], 60.00th=[33162], 00:40:02.469 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34341], 00:40:02.469 | 99.00th=[35390], 99.50th=[35914], 99.90th=[54264], 99.95th=[54264], 00:40:02.469 | 99.99th=[54264] 00:40:02.469 bw ( KiB/s): min= 1792, max= 2048, per=4.11%, avg=1926.40, stdev=77.42, samples=20 00:40:02.469 iops : min= 448, max= 512, avg=481.60, stdev=19.35, samples=20 00:40:02.469 lat (msec) : 20=0.66%, 50=99.01%, 100=0.33% 00:40:02.469 cpu : usr=99.00%, sys=0.67%, ctx=13, majf=0, minf=9 00:40:02.469 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:40:02.469 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 issued rwts: total=4832,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.469 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.469 filename2: (groupid=0, jobs=1): err= 0: pid=2198808: Wed Nov 6 09:15:51 2024 00:40:02.469 read: IOPS=494, BW=1979KiB/s (2027kB/s)(19.4MiB/10047msec) 00:40:02.469 slat (usec): min=5, max=110, avg=21.30, stdev=16.01 00:40:02.469 clat (usec): min=12154, max=53983, avg=32107.16, stdev=4276.01 00:40:02.469 lat (usec): min=12161, max=53999, avg=32128.46, stdev=4278.14 00:40:02.469 clat percentiles (usec): 00:40:02.469 | 1.00th=[19268], 5.00th=[23200], 10.00th=[27395], 20.00th=[32113], 00:40:02.469 | 30.00th=[32375], 40.00th=[32375], 50.00th=[32637], 60.00th=[32900], 00:40:02.469 | 70.00th=[33162], 80.00th=[33424], 90.00th=[33817], 95.00th=[34866], 00:40:02.469 | 99.00th=[48497], 99.50th=[53740], 99.90th=[53740], 99.95th=[53740], 00:40:02.469 | 99.99th=[53740] 00:40:02.469 bw ( KiB/s): min= 1795, max= 2288, per=4.23%, avg=1982.55, stdev=113.22, samples=20 00:40:02.469 iops : min= 448, max= 572, avg=495.60, stdev=28.37, samples=20 00:40:02.469 lat (msec) : 20=1.77%, 50=97.47%, 100=0.76% 00:40:02.469 cpu : usr=98.98%, sys=0.69%, ctx=14, majf=0, minf=9 00:40:02.469 IO depths : 1=4.2%, 2=8.7%, 4=19.0%, 8=59.1%, 16=9.0%, 32=0.0%, >=64=0.0% 00:40:02.469 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 complete : 0=0.0%, 4=92.5%, 8=2.4%, 16=5.1%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:02.469 issued rwts: total=4972,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:02.469 latency : target=0, window=0, percentile=100.00%, depth=16 00:40:02.469 00:40:02.469 Run status group 0 (all jobs): 00:40:02.469 READ: bw=45.8MiB/s (48.0MB/s), 1931KiB/s-2026KiB/s (1977kB/s-2075kB/s), io=460MiB (482MB), run=10001-10047msec 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@113 -- # destroy_subsystems 0 1 2 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 1 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=1 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.469 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 2 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=2 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null2 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # NULL_DIF=1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # bs=8k,16k,128k 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # numjobs=2 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # iodepth=8 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # runtime=5 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # files=1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@117 -- # create_subsystems 0 1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.470 bdev_null0 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.470 [2024-11-06 09:15:51.278798] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.470 bdev_null1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@118 -- # fio /dev/fd/62 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@118 -- # create_json_sub_conf 0 1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # config=() 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # local subsystem config 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:40:02.470 { 00:40:02.470 "params": { 00:40:02.470 "name": "Nvme$subsystem", 00:40:02.470 "trtype": "$TEST_TRANSPORT", 00:40:02.470 "traddr": "$NVMF_FIRST_TARGET_IP", 00:40:02.470 "adrfam": "ipv4", 00:40:02.470 "trsvcid": "$NVMF_PORT", 00:40:02.470 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:40:02.470 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:40:02.470 "hdgst": ${hdgst:-false}, 00:40:02.470 "ddgst": ${ddgst:-false} 00:40:02.470 }, 00:40:02.470 "method": "bdev_nvme_attach_controller" 00:40:02.470 } 00:40:02.470 EOF 00:40:02.470 )") 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # local sanitizers 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # shift 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local asan_lib= 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libasan 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:40:02.470 { 00:40:02.470 "params": { 00:40:02.470 "name": "Nvme$subsystem", 00:40:02.470 "trtype": "$TEST_TRANSPORT", 00:40:02.470 "traddr": "$NVMF_FIRST_TARGET_IP", 00:40:02.470 "adrfam": "ipv4", 00:40:02.470 "trsvcid": "$NVMF_PORT", 00:40:02.470 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:40:02.470 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:40:02.470 "hdgst": ${hdgst:-false}, 00:40:02.470 "ddgst": ${ddgst:-false} 00:40:02.470 }, 00:40:02.470 "method": "bdev_nvme_attach_controller" 00:40:02.470 } 00:40:02.470 EOF 00:40:02.470 )") 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@580 -- # cat 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@582 -- # jq . 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@583 -- # IFS=, 00:40:02.470 09:15:51 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:40:02.470 "params": { 00:40:02.470 "name": "Nvme0", 00:40:02.470 "trtype": "tcp", 00:40:02.470 "traddr": "10.0.0.2", 00:40:02.470 "adrfam": "ipv4", 00:40:02.470 "trsvcid": "4420", 00:40:02.470 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:40:02.470 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:40:02.470 "hdgst": false, 00:40:02.470 "ddgst": false 00:40:02.470 }, 00:40:02.470 "method": "bdev_nvme_attach_controller" 00:40:02.470 },{ 00:40:02.470 "params": { 00:40:02.470 "name": "Nvme1", 00:40:02.471 "trtype": "tcp", 00:40:02.471 "traddr": "10.0.0.2", 00:40:02.471 "adrfam": "ipv4", 00:40:02.471 "trsvcid": "4420", 00:40:02.471 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:40:02.471 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:40:02.471 "hdgst": false, 00:40:02.471 "ddgst": false 00:40:02.471 }, 00:40:02.471 "method": "bdev_nvme_attach_controller" 00:40:02.471 }' 00:40:02.471 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:40:02.471 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:40:02.471 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:40:02.471 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:40:02.471 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:40:02.471 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:40:02.471 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:40:02.471 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:40:02.471 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:40:02.471 09:15:51 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:40:02.471 filename0: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:40:02.471 ... 00:40:02.471 filename1: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:40:02.471 ... 00:40:02.471 fio-3.35 00:40:02.471 Starting 4 threads 00:40:07.760 00:40:07.761 filename0: (groupid=0, jobs=1): err= 0: pid=2201086: Wed Nov 6 09:15:57 2024 00:40:07.761 read: IOPS=2118, BW=16.5MiB/s (17.4MB/s)(82.8MiB/5002msec) 00:40:07.761 slat (nsec): min=5418, max=99488, avg=6806.89, stdev=3531.87 00:40:07.761 clat (usec): min=2113, max=6443, avg=3757.84, stdev=700.22 00:40:07.761 lat (usec): min=2120, max=6449, avg=3764.65, stdev=700.06 00:40:07.761 clat percentiles (usec): 00:40:07.761 | 1.00th=[ 2638], 5.00th=[ 2900], 10.00th=[ 3064], 20.00th=[ 3294], 00:40:07.761 | 30.00th=[ 3425], 40.00th=[ 3523], 50.00th=[ 3589], 60.00th=[ 3720], 00:40:07.761 | 70.00th=[ 3785], 80.00th=[ 4015], 90.00th=[ 5014], 95.00th=[ 5342], 00:40:07.761 | 99.00th=[ 5866], 99.50th=[ 5932], 99.90th=[ 6259], 99.95th=[ 6390], 00:40:07.761 | 99.99th=[ 6456] 00:40:07.761 bw ( KiB/s): min=16416, max=17888, per=25.72%, avg=16963.56, stdev=512.36, samples=9 00:40:07.761 iops : min= 2052, max= 2236, avg=2120.44, stdev=64.05, samples=9 00:40:07.761 lat (msec) : 4=79.76%, 10=20.24% 00:40:07.761 cpu : usr=97.82%, sys=1.92%, ctx=18, majf=0, minf=95 00:40:07.761 IO depths : 1=0.1%, 2=1.0%, 4=71.1%, 8=27.9%, 16=0.0%, 32=0.0%, >=64=0.0% 00:40:07.761 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:07.761 complete : 0=0.0%, 4=93.2%, 8=6.8%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:07.761 issued rwts: total=10596,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:07.761 latency : target=0, window=0, percentile=100.00%, depth=8 00:40:07.761 filename0: (groupid=0, jobs=1): err= 0: pid=2201087: Wed Nov 6 09:15:57 2024 00:40:07.761 read: IOPS=2099, BW=16.4MiB/s (17.2MB/s)(82.0MiB/5002msec) 00:40:07.761 slat (nsec): min=5419, max=62443, avg=6623.44, stdev=2851.39 00:40:07.761 clat (usec): min=1139, max=6462, avg=3792.36, stdev=720.51 00:40:07.761 lat (usec): min=1160, max=6468, avg=3798.98, stdev=720.21 00:40:07.761 clat percentiles (usec): 00:40:07.761 | 1.00th=[ 2442], 5.00th=[ 2900], 10.00th=[ 3130], 20.00th=[ 3359], 00:40:07.761 | 30.00th=[ 3490], 40.00th=[ 3556], 50.00th=[ 3621], 60.00th=[ 3752], 00:40:07.761 | 70.00th=[ 3851], 80.00th=[ 4047], 90.00th=[ 5014], 95.00th=[ 5342], 00:40:07.761 | 99.00th=[ 5866], 99.50th=[ 5932], 99.90th=[ 6325], 99.95th=[ 6456], 00:40:07.761 | 99.99th=[ 6456] 00:40:07.761 bw ( KiB/s): min=16352, max=17184, per=25.42%, avg=16771.56, stdev=281.91, samples=9 00:40:07.761 iops : min= 2044, max= 2148, avg=2096.44, stdev=35.24, samples=9 00:40:07.761 lat (msec) : 2=0.55%, 4=77.89%, 10=21.56% 00:40:07.761 cpu : usr=96.84%, sys=2.92%, ctx=7, majf=0, minf=114 00:40:07.761 IO depths : 1=0.1%, 2=0.5%, 4=71.5%, 8=28.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:40:07.761 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:07.761 complete : 0=0.0%, 4=93.3%, 8=6.7%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:07.761 issued rwts: total=10502,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:07.761 latency : target=0, window=0, percentile=100.00%, depth=8 00:40:07.761 filename1: (groupid=0, jobs=1): err= 0: pid=2201088: Wed Nov 6 09:15:57 2024 00:40:07.761 read: IOPS=2043, BW=16.0MiB/s (16.7MB/s)(79.9MiB/5002msec) 00:40:07.761 slat (nsec): min=5416, max=99954, avg=7287.54, stdev=3659.93 00:40:07.761 clat (usec): min=1869, max=6856, avg=3895.05, stdev=671.67 00:40:07.761 lat (usec): min=1877, max=6886, avg=3902.33, stdev=671.68 00:40:07.761 clat percentiles (usec): 00:40:07.761 | 1.00th=[ 2835], 5.00th=[ 3228], 10.00th=[ 3326], 20.00th=[ 3490], 00:40:07.761 | 30.00th=[ 3556], 40.00th=[ 3621], 50.00th=[ 3720], 60.00th=[ 3818], 00:40:07.761 | 70.00th=[ 3916], 80.00th=[ 4080], 90.00th=[ 5211], 95.00th=[ 5538], 00:40:07.761 | 99.00th=[ 5932], 99.50th=[ 6194], 99.90th=[ 6521], 99.95th=[ 6587], 00:40:07.761 | 99.99th=[ 6783] 00:40:07.761 bw ( KiB/s): min=15504, max=17248, per=24.83%, avg=16376.89, stdev=580.71, samples=9 00:40:07.761 iops : min= 1938, max= 2156, avg=2047.11, stdev=72.59, samples=9 00:40:07.761 lat (msec) : 2=0.01%, 4=76.28%, 10=23.71% 00:40:07.761 cpu : usr=97.14%, sys=2.62%, ctx=7, majf=0, minf=92 00:40:07.761 IO depths : 1=0.1%, 2=0.2%, 4=72.4%, 8=27.3%, 16=0.0%, 32=0.0%, >=64=0.0% 00:40:07.761 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:07.761 complete : 0=0.0%, 4=92.5%, 8=7.5%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:07.761 issued rwts: total=10221,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:07.761 latency : target=0, window=0, percentile=100.00%, depth=8 00:40:07.761 filename1: (groupid=0, jobs=1): err= 0: pid=2201089: Wed Nov 6 09:15:57 2024 00:40:07.761 read: IOPS=1984, BW=15.5MiB/s (16.3MB/s)(77.5MiB/5002msec) 00:40:07.761 slat (nsec): min=5414, max=60965, avg=6388.45, stdev=2527.86 00:40:07.761 clat (usec): min=1790, max=8777, avg=4013.65, stdev=733.55 00:40:07.761 lat (usec): min=1796, max=8804, avg=4020.04, stdev=733.45 00:40:07.761 clat percentiles (usec): 00:40:07.761 | 1.00th=[ 3163], 5.00th=[ 3294], 10.00th=[ 3425], 20.00th=[ 3523], 00:40:07.761 | 30.00th=[ 3589], 40.00th=[ 3687], 50.00th=[ 3818], 60.00th=[ 3851], 00:40:07.761 | 70.00th=[ 3982], 80.00th=[ 4228], 90.00th=[ 5342], 95.00th=[ 5669], 00:40:07.761 | 99.00th=[ 6128], 99.50th=[ 6259], 99.90th=[ 6718], 99.95th=[ 8455], 00:40:07.761 | 99.99th=[ 8717] 00:40:07.761 bw ( KiB/s): min=15488, max=16080, per=24.01%, avg=15841.67, stdev=213.73, samples=9 00:40:07.761 iops : min= 1936, max= 2010, avg=1980.11, stdev=26.89, samples=9 00:40:07.761 lat (msec) : 2=0.03%, 4=71.20%, 10=28.77% 00:40:07.761 cpu : usr=97.36%, sys=2.36%, ctx=5, majf=0, minf=90 00:40:07.761 IO depths : 1=0.1%, 2=0.1%, 4=72.9%, 8=27.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:40:07.761 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:07.761 complete : 0=0.0%, 4=92.5%, 8=7.5%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:07.761 issued rwts: total=9925,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:07.761 latency : target=0, window=0, percentile=100.00%, depth=8 00:40:07.761 00:40:07.761 Run status group 0 (all jobs): 00:40:07.761 READ: bw=64.4MiB/s (67.5MB/s), 15.5MiB/s-16.5MiB/s (16.3MB/s-17.4MB/s), io=322MiB (338MB), run=5002-5002msec 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@119 -- # destroy_subsystems 0 1 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 1 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=1 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:07.761 00:40:07.761 real 0m24.566s 00:40:07.761 user 5m20.997s 00:40:07.761 sys 0m4.273s 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1126 -- # xtrace_disable 00:40:07.761 09:15:57 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:40:07.761 ************************************ 00:40:07.761 END TEST fio_dif_rand_params 00:40:07.761 ************************************ 00:40:07.761 09:15:57 nvmf_dif -- target/dif.sh@144 -- # run_test fio_dif_digest fio_dif_digest 00:40:07.761 09:15:57 nvmf_dif -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:40:07.761 09:15:57 nvmf_dif -- common/autotest_common.sh@1107 -- # xtrace_disable 00:40:07.761 09:15:57 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:40:07.761 ************************************ 00:40:07.761 START TEST fio_dif_digest 00:40:07.761 ************************************ 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1125 -- # fio_dif_digest 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@123 -- # local NULL_DIF 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@124 -- # local bs numjobs runtime iodepth files 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@125 -- # local hdgst ddgst 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # NULL_DIF=3 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # bs=128k,128k,128k 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # numjobs=3 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # iodepth=3 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # runtime=10 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@128 -- # hdgst=true 00:40:07.761 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@128 -- # ddgst=true 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@130 -- # create_subsystems 0 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@28 -- # local sub 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@30 -- # for sub in "$@" 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@31 -- # create_subsystem 0 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@18 -- # local sub_id=0 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:40:07.762 bdev_null0 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:40:07.762 [2024-11-06 09:15:57.666618] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@131 -- # fio /dev/fd/62 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@131 -- # create_json_sub_conf 0 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- nvmf/common.sh@558 -- # config=() 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- nvmf/common.sh@558 -- # local subsystem config 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- nvmf/common.sh@560 -- # for subsystem in "${@:-1}" 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- nvmf/common.sh@580 -- # config+=("$(cat <<-EOF 00:40:07.762 { 00:40:07.762 "params": { 00:40:07.762 "name": "Nvme$subsystem", 00:40:07.762 "trtype": "$TEST_TRANSPORT", 00:40:07.762 "traddr": "$NVMF_FIRST_TARGET_IP", 00:40:07.762 "adrfam": "ipv4", 00:40:07.762 "trsvcid": "$NVMF_PORT", 00:40:07.762 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:40:07.762 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:40:07.762 "hdgst": ${hdgst:-false}, 00:40:07.762 "ddgst": ${ddgst:-false} 00:40:07.762 }, 00:40:07.762 "method": "bdev_nvme_attach_controller" 00:40:07.762 } 00:40:07.762 EOF 00:40:07.762 )") 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@82 -- # gen_fio_conf 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@54 -- # local file 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1339 -- # local sanitizers 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@56 -- # cat 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1341 -- # shift 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1343 -- # local asan_lib= 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- nvmf/common.sh@580 -- # cat 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@72 -- # (( file = 1 )) 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # grep libasan 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- target/dif.sh@72 -- # (( file <= files )) 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- nvmf/common.sh@582 -- # jq . 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- nvmf/common.sh@583 -- # IFS=, 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- nvmf/common.sh@584 -- # printf '%s\n' '{ 00:40:07.762 "params": { 00:40:07.762 "name": "Nvme0", 00:40:07.762 "trtype": "tcp", 00:40:07.762 "traddr": "10.0.0.2", 00:40:07.762 "adrfam": "ipv4", 00:40:07.762 "trsvcid": "4420", 00:40:07.762 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:40:07.762 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:40:07.762 "hdgst": true, 00:40:07.762 "ddgst": true 00:40:07.762 }, 00:40:07.762 "method": "bdev_nvme_attach_controller" 00:40:07.762 }' 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # asan_lib= 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # asan_lib= 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:40:07.762 09:15:57 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:40:08.022 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:40:08.022 ... 00:40:08.022 fio-3.35 00:40:08.022 Starting 3 threads 00:40:20.263 00:40:20.263 filename0: (groupid=0, jobs=1): err= 0: pid=2202459: Wed Nov 6 09:16:08 2024 00:40:20.263 read: IOPS=233, BW=29.1MiB/s (30.5MB/s)(293MiB/10047msec) 00:40:20.263 slat (nsec): min=5665, max=48312, avg=7120.66, stdev=1782.86 00:40:20.263 clat (usec): min=7252, max=49333, avg=12847.83, stdev=1984.23 00:40:20.263 lat (usec): min=7258, max=49339, avg=12854.95, stdev=1984.19 00:40:20.263 clat percentiles (usec): 00:40:20.263 | 1.00th=[ 8160], 5.00th=[ 9110], 10.00th=[ 9896], 20.00th=[11863], 00:40:20.263 | 30.00th=[12387], 40.00th=[12780], 50.00th=[13173], 60.00th=[13435], 00:40:20.263 | 70.00th=[13698], 80.00th=[14091], 90.00th=[14484], 95.00th=[15008], 00:40:20.263 | 99.00th=[16057], 99.50th=[16581], 99.90th=[17433], 99.95th=[46400], 00:40:20.263 | 99.99th=[49546] 00:40:20.263 bw ( KiB/s): min=27136, max=33792, per=36.90%, avg=29926.40, stdev=1720.11, samples=20 00:40:20.263 iops : min= 212, max= 264, avg=233.80, stdev=13.44, samples=20 00:40:20.263 lat (msec) : 10=10.29%, 20=89.62%, 50=0.09% 00:40:20.263 cpu : usr=95.81%, sys=3.94%, ctx=25, majf=0, minf=169 00:40:20.263 IO depths : 1=0.1%, 2=100.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:40:20.263 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:20.263 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:20.263 issued rwts: total=2341,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:20.263 latency : target=0, window=0, percentile=100.00%, depth=3 00:40:20.263 filename0: (groupid=0, jobs=1): err= 0: pid=2202460: Wed Nov 6 09:16:08 2024 00:40:20.263 read: IOPS=218, BW=27.3MiB/s (28.6MB/s)(274MiB/10046msec) 00:40:20.263 slat (nsec): min=5724, max=37526, avg=6794.52, stdev=1306.29 00:40:20.263 clat (usec): min=8587, max=56868, avg=13701.78, stdev=2784.02 00:40:20.263 lat (usec): min=8594, max=56875, avg=13708.58, stdev=2784.08 00:40:20.263 clat percentiles (usec): 00:40:20.263 | 1.00th=[ 9765], 5.00th=[10814], 10.00th=[11600], 20.00th=[12649], 00:40:20.263 | 30.00th=[13042], 40.00th=[13435], 50.00th=[13698], 60.00th=[13960], 00:40:20.263 | 70.00th=[14222], 80.00th=[14615], 90.00th=[15139], 95.00th=[15664], 00:40:20.263 | 99.00th=[16450], 99.50th=[17171], 99.90th=[55837], 99.95th=[56361], 00:40:20.263 | 99.99th=[56886] 00:40:20.263 bw ( KiB/s): min=24576, max=30976, per=34.62%, avg=28070.40, stdev=1380.54, samples=20 00:40:20.263 iops : min= 192, max= 242, avg=219.30, stdev=10.79, samples=20 00:40:20.263 lat (msec) : 10=1.41%, 20=98.22%, 50=0.09%, 100=0.27% 00:40:20.263 cpu : usr=94.29%, sys=5.46%, ctx=23, majf=0, minf=118 00:40:20.263 IO depths : 1=0.1%, 2=100.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:40:20.263 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:20.263 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:20.263 issued rwts: total=2195,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:20.263 latency : target=0, window=0, percentile=100.00%, depth=3 00:40:20.263 filename0: (groupid=0, jobs=1): err= 0: pid=2202461: Wed Nov 6 09:16:08 2024 00:40:20.263 read: IOPS=182, BW=22.8MiB/s (23.9MB/s)(229MiB/10037msec) 00:40:20.263 slat (nsec): min=5647, max=31156, avg=7085.23, stdev=1500.65 00:40:20.263 clat (usec): min=9180, max=97837, avg=16450.38, stdev=7728.40 00:40:20.263 lat (usec): min=9186, max=97846, avg=16457.47, stdev=7728.47 00:40:20.263 clat percentiles (usec): 00:40:20.263 | 1.00th=[11994], 5.00th=[13173], 10.00th=[13698], 20.00th=[14222], 00:40:20.263 | 30.00th=[14484], 40.00th=[14877], 50.00th=[15139], 60.00th=[15401], 00:40:20.263 | 70.00th=[15795], 80.00th=[16319], 90.00th=[16909], 95.00th=[17695], 00:40:20.263 | 99.00th=[56886], 99.50th=[57410], 99.90th=[95945], 99.95th=[98042], 00:40:20.263 | 99.99th=[98042] 00:40:20.263 bw ( KiB/s): min=16128, max=25600, per=28.82%, avg=23372.80, stdev=2414.53, samples=20 00:40:20.263 iops : min= 126, max= 200, avg=182.60, stdev=18.86, samples=20 00:40:20.263 lat (msec) : 10=0.16%, 20=96.72%, 50=0.05%, 100=3.06% 00:40:20.263 cpu : usr=95.30%, sys=4.41%, ctx=19, majf=0, minf=122 00:40:20.263 IO depths : 1=0.1%, 2=99.9%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:40:20.263 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:20.263 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:40:20.263 issued rwts: total=1829,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:40:20.263 latency : target=0, window=0, percentile=100.00%, depth=3 00:40:20.263 00:40:20.263 Run status group 0 (all jobs): 00:40:20.263 READ: bw=79.2MiB/s (83.0MB/s), 22.8MiB/s-29.1MiB/s (23.9MB/s-30.5MB/s), io=796MiB (834MB), run=10037-10047msec 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- target/dif.sh@132 -- # destroy_subsystems 0 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- target/dif.sh@43 -- # local sub 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- target/dif.sh@45 -- # for sub in "$@" 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- target/dif.sh@46 -- # destroy_subsystem 0 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- target/dif.sh@36 -- # local sub_id=0 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:20.263 00:40:20.263 real 0m11.339s 00:40:20.263 user 0m46.618s 00:40:20.263 sys 0m1.725s 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1126 -- # xtrace_disable 00:40:20.263 09:16:08 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:40:20.263 ************************************ 00:40:20.263 END TEST fio_dif_digest 00:40:20.263 ************************************ 00:40:20.263 09:16:09 nvmf_dif -- target/dif.sh@146 -- # trap - SIGINT SIGTERM EXIT 00:40:20.263 09:16:09 nvmf_dif -- target/dif.sh@147 -- # nvmftestfini 00:40:20.263 09:16:09 nvmf_dif -- nvmf/common.sh@514 -- # nvmfcleanup 00:40:20.263 09:16:09 nvmf_dif -- nvmf/common.sh@121 -- # sync 00:40:20.263 09:16:09 nvmf_dif -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:40:20.264 09:16:09 nvmf_dif -- nvmf/common.sh@124 -- # set +e 00:40:20.264 09:16:09 nvmf_dif -- nvmf/common.sh@125 -- # for i in {1..20} 00:40:20.264 09:16:09 nvmf_dif -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:40:20.264 rmmod nvme_tcp 00:40:20.264 rmmod nvme_fabrics 00:40:20.264 rmmod nvme_keyring 00:40:20.264 09:16:09 nvmf_dif -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:40:20.264 09:16:09 nvmf_dif -- nvmf/common.sh@128 -- # set -e 00:40:20.264 09:16:09 nvmf_dif -- nvmf/common.sh@129 -- # return 0 00:40:20.264 09:16:09 nvmf_dif -- nvmf/common.sh@515 -- # '[' -n 2191868 ']' 00:40:20.264 09:16:09 nvmf_dif -- nvmf/common.sh@516 -- # killprocess 2191868 00:40:20.264 09:16:09 nvmf_dif -- common/autotest_common.sh@950 -- # '[' -z 2191868 ']' 00:40:20.264 09:16:09 nvmf_dif -- common/autotest_common.sh@954 -- # kill -0 2191868 00:40:20.264 09:16:09 nvmf_dif -- common/autotest_common.sh@955 -- # uname 00:40:20.264 09:16:09 nvmf_dif -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:40:20.264 09:16:09 nvmf_dif -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2191868 00:40:20.264 09:16:09 nvmf_dif -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:40:20.264 09:16:09 nvmf_dif -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:40:20.264 09:16:09 nvmf_dif -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2191868' 00:40:20.264 killing process with pid 2191868 00:40:20.264 09:16:09 nvmf_dif -- common/autotest_common.sh@969 -- # kill 2191868 00:40:20.264 09:16:09 nvmf_dif -- common/autotest_common.sh@974 -- # wait 2191868 00:40:20.264 09:16:09 nvmf_dif -- nvmf/common.sh@518 -- # '[' iso == iso ']' 00:40:20.264 09:16:09 nvmf_dif -- nvmf/common.sh@519 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:40:22.812 Waiting for block devices as requested 00:40:22.812 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:40:22.812 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:40:22.812 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:40:22.812 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:40:22.812 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:40:23.072 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:40:23.072 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:40:23.072 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:40:23.333 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:40:23.333 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:40:23.333 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:40:23.594 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:40:23.594 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:40:23.594 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:40:23.854 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:40:23.854 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:40:23.854 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:40:24.114 09:16:14 nvmf_dif -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:40:24.114 09:16:14 nvmf_dif -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:40:24.114 09:16:14 nvmf_dif -- nvmf/common.sh@297 -- # iptr 00:40:24.114 09:16:14 nvmf_dif -- nvmf/common.sh@789 -- # iptables-save 00:40:24.114 09:16:14 nvmf_dif -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:40:24.114 09:16:14 nvmf_dif -- nvmf/common.sh@789 -- # iptables-restore 00:40:24.114 09:16:14 nvmf_dif -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:40:24.114 09:16:14 nvmf_dif -- nvmf/common.sh@302 -- # remove_spdk_ns 00:40:24.115 09:16:14 nvmf_dif -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:40:24.115 09:16:14 nvmf_dif -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:40:24.115 09:16:14 nvmf_dif -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:40:26.657 09:16:16 nvmf_dif -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:40:26.657 00:40:26.657 real 1m17.926s 00:40:26.657 user 8m10.159s 00:40:26.657 sys 0m21.092s 00:40:26.657 09:16:16 nvmf_dif -- common/autotest_common.sh@1126 -- # xtrace_disable 00:40:26.657 09:16:16 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:40:26.657 ************************************ 00:40:26.657 END TEST nvmf_dif 00:40:26.657 ************************************ 00:40:26.657 09:16:16 -- spdk/autotest.sh@286 -- # run_test nvmf_abort_qd_sizes /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:40:26.657 09:16:16 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:40:26.657 09:16:16 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:40:26.657 09:16:16 -- common/autotest_common.sh@10 -- # set +x 00:40:26.657 ************************************ 00:40:26.657 START TEST nvmf_abort_qd_sizes 00:40:26.657 ************************************ 00:40:26.657 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:40:26.657 * Looking for test storage... 00:40:26.657 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:40:26.657 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:40:26.657 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@1689 -- # lcov --version 00:40:26.657 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:40:26.657 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:40:26.657 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:40:26.657 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@333 -- # local ver1 ver1_l 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@334 -- # local ver2 ver2_l 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@336 -- # IFS=.-: 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@336 -- # read -ra ver1 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@337 -- # IFS=.-: 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@337 -- # read -ra ver2 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@338 -- # local 'op=<' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@340 -- # ver1_l=2 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@341 -- # ver2_l=1 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@344 -- # case "$op" in 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@345 -- # : 1 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@364 -- # (( v = 0 )) 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@365 -- # decimal 1 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@353 -- # local d=1 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@355 -- # echo 1 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@365 -- # ver1[v]=1 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@366 -- # decimal 2 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@353 -- # local d=2 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@355 -- # echo 2 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@366 -- # ver2[v]=2 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@368 -- # return 0 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:40:26.658 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:26.658 --rc genhtml_branch_coverage=1 00:40:26.658 --rc genhtml_function_coverage=1 00:40:26.658 --rc genhtml_legend=1 00:40:26.658 --rc geninfo_all_blocks=1 00:40:26.658 --rc geninfo_unexecuted_blocks=1 00:40:26.658 00:40:26.658 ' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:40:26.658 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:26.658 --rc genhtml_branch_coverage=1 00:40:26.658 --rc genhtml_function_coverage=1 00:40:26.658 --rc genhtml_legend=1 00:40:26.658 --rc geninfo_all_blocks=1 00:40:26.658 --rc geninfo_unexecuted_blocks=1 00:40:26.658 00:40:26.658 ' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:40:26.658 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:26.658 --rc genhtml_branch_coverage=1 00:40:26.658 --rc genhtml_function_coverage=1 00:40:26.658 --rc genhtml_legend=1 00:40:26.658 --rc geninfo_all_blocks=1 00:40:26.658 --rc geninfo_unexecuted_blocks=1 00:40:26.658 00:40:26.658 ' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:40:26.658 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:40:26.658 --rc genhtml_branch_coverage=1 00:40:26.658 --rc genhtml_function_coverage=1 00:40:26.658 --rc genhtml_legend=1 00:40:26.658 --rc geninfo_all_blocks=1 00:40:26.658 --rc geninfo_unexecuted_blocks=1 00:40:26.658 00:40:26.658 ' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@7 -- # uname -s 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@15 -- # shopt -s extglob 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- paths/export.sh@5 -- # export PATH 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@51 -- # : 0 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:40:26.658 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@55 -- # have_pci_nics=0 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@70 -- # nvmftestinit 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@467 -- # '[' -z tcp ']' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@472 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@474 -- # prepare_net_devs 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@436 -- # local -g is_hw=no 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@438 -- # remove_spdk_ns 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@440 -- # [[ phy != virt ]] 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@440 -- # gather_supported_nvmf_pci_devs 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- nvmf/common.sh@309 -- # xtrace_disable 00:40:26.658 09:16:16 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@313 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@315 -- # pci_devs=() 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@315 -- # local -a pci_devs 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@316 -- # pci_net_devs=() 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@316 -- # local -a pci_net_devs 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@317 -- # pci_drivers=() 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@317 -- # local -A pci_drivers 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@319 -- # net_devs=() 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@319 -- # local -ga net_devs 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@320 -- # e810=() 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@320 -- # local -ga e810 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@321 -- # x722=() 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@321 -- # local -ga x722 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@322 -- # mlx=() 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@322 -- # local -ga mlx 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@325 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@326 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@328 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@330 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@332 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@334 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@336 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@338 -- # mlx+=(${pci_bus_cache["$mellanox:0x101b"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@340 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@341 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@343 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@344 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@346 -- # pci_devs+=("${e810[@]}") 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@347 -- # [[ tcp == rdma ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@353 -- # [[ e810 == mlx5 ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@355 -- # [[ e810 == e810 ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@356 -- # pci_devs=("${e810[@]}") 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@361 -- # (( 2 == 0 )) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.0 (0x8086 - 0x159b)' 00:40:34.796 Found 0000:4b:00.0 (0x8086 - 0x159b) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@366 -- # for pci in "${pci_devs[@]}" 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@367 -- # echo 'Found 0000:4b:00.1 (0x8086 - 0x159b)' 00:40:34.796 Found 0000:4b:00.1 (0x8086 - 0x159b) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@368 -- # [[ ice == unknown ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@372 -- # [[ ice == unbound ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@376 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@377 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@378 -- # [[ tcp == rdma ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@392 -- # (( 0 > 0 )) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@398 -- # [[ e810 == e810 ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@398 -- # [[ tcp == rdma ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@416 -- # [[ up == up ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.0: cvl_0_0' 00:40:34.796 Found net devices under 0000:4b:00.0: cvl_0_0 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@408 -- # for pci in "${pci_devs[@]}" 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@409 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@414 -- # [[ tcp == tcp ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@415 -- # for net_dev in "${!pci_net_devs[@]}" 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@416 -- # [[ up == up ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@420 -- # (( 1 == 0 )) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@425 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@426 -- # echo 'Found net devices under 0000:4b:00.1: cvl_0_1' 00:40:34.796 Found net devices under 0000:4b:00.1: cvl_0_1 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@427 -- # net_devs+=("${pci_net_devs[@]}") 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@430 -- # (( 2 == 0 )) 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@440 -- # is_hw=yes 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@442 -- # [[ yes == yes ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@443 -- # [[ tcp == tcp ]] 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@444 -- # nvmf_tcp_init 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@250 -- # NVMF_FIRST_INITIATOR_IP=10.0.0.1 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@251 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:40:34.796 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@252 -- # NVMF_INITIATOR_IP=10.0.0.1 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@253 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@256 -- # (( 2 > 1 )) 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@258 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@259 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@262 -- # NVMF_SECOND_TARGET_IP= 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@263 -- # NVMF_SECOND_INITIATOR_IP= 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@265 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@266 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@267 -- # ip -4 addr flush cvl_0_0 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@268 -- # ip -4 addr flush cvl_0_1 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@271 -- # ip netns add cvl_0_0_ns_spdk 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@274 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@277 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@278 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@281 -- # ip link set cvl_0_1 up 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@283 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@284 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@287 -- # ipts -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@788 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT -m comment --comment 'SPDK_NVMF:-I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT' 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@290 -- # ping -c 1 10.0.0.2 00:40:34.797 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:40:34.797 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.664 ms 00:40:34.797 00:40:34.797 --- 10.0.0.2 ping statistics --- 00:40:34.797 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:40:34.797 rtt min/avg/max/mdev = 0.664/0.664/0.664/0.000 ms 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@291 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:40:34.797 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:40:34.797 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.312 ms 00:40:34.797 00:40:34.797 --- 10.0.0.1 ping statistics --- 00:40:34.797 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:40:34.797 rtt min/avg/max/mdev = 0.312/0.312/0.312/0.000 ms 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@293 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@448 -- # return 0 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@476 -- # '[' iso == iso ']' 00:40:34.797 09:16:23 nvmf_abort_qd_sizes -- nvmf/common.sh@477 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:40:37.342 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:40:37.342 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:40:37.342 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:40:37.342 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:40:37.342 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:40:37.342 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:40:37.342 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:40:37.343 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:40:37.343 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:40:37.343 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:40:37.343 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:40:37.343 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:40:37.343 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:40:37.343 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:40:37.343 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:40:37.343 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:40:37.343 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:40:37.603 09:16:27 nvmf_abort_qd_sizes -- nvmf/common.sh@480 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- nvmf/common.sh@481 -- # [[ tcp == \r\d\m\a ]] 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- nvmf/common.sh@490 -- # [[ tcp == \t\c\p ]] 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- nvmf/common.sh@491 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- nvmf/common.sh@494 -- # '[' tcp == tcp ']' 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- nvmf/common.sh@500 -- # modprobe nvme-tcp 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@71 -- # nvmfappstart -m 0xf 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- nvmf/common.sh@505 -- # timing_enter start_nvmf_tgt 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@724 -- # xtrace_disable 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- nvmf/common.sh@507 -- # nvmfpid=2211875 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- nvmf/common.sh@508 -- # waitforlisten 2211875 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- nvmf/common.sh@506 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xf 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@831 -- # '[' -z 2211875 ']' 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@836 -- # local max_retries=100 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:40:37.604 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@840 -- # xtrace_disable 00:40:37.604 09:16:27 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:40:37.604 [2024-11-06 09:16:27.650723] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:40:37.604 [2024-11-06 09:16:27.650795] [ DPDK EAL parameters: nvmf -c 0xf --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:40:37.865 [2024-11-06 09:16:27.731919] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 4 00:40:37.865 [2024-11-06 09:16:27.772340] app.c: 612:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:40:37.865 [2024-11-06 09:16:27.772376] app.c: 613:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:40:37.865 [2024-11-06 09:16:27.772384] app.c: 618:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:40:37.865 [2024-11-06 09:16:27.772391] app.c: 619:app_setup_trace: *NOTICE*: SPDK application currently running. 00:40:37.865 [2024-11-06 09:16:27.772397] app.c: 620:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:40:37.865 [2024-11-06 09:16:27.774167] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:40:37.865 [2024-11-06 09:16:27.774307] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 2 00:40:37.865 [2024-11-06 09:16:27.774382] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:40:37.865 [2024-11-06 09:16:27.774383] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 3 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- common/autotest_common.sh@864 -- # return 0 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- nvmf/common.sh@509 -- # timing_exit start_nvmf_tgt 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- common/autotest_common.sh@730 -- # xtrace_disable 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- nvmf/common.sh@510 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@73 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini || :; clean_kernel_target' SIGINT SIGTERM EXIT 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@75 -- # mapfile -t nvmes 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@75 -- # nvme_in_userspace 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@312 -- # local bdf bdfs 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@313 -- # local nvmes 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@315 -- # [[ -n 0000:65:00.0 ]] 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@316 -- # nvmes=(${pci_bus_cache["0x010802"]}) 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@321 -- # for bdf in "${nvmes[@]}" 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@322 -- # [[ -e /sys/bus/pci/drivers/nvme/0000:65:00.0 ]] 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@323 -- # uname -s 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@323 -- # [[ Linux == FreeBSD ]] 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@326 -- # bdfs+=("$bdf") 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@328 -- # (( 1 )) 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- scripts/common.sh@329 -- # printf '%s\n' 0000:65:00.0 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@76 -- # (( 1 > 0 )) 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@78 -- # nvme=0000:65:00.0 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@80 -- # run_test spdk_target_abort spdk_target 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- common/autotest_common.sh@1107 -- # xtrace_disable 00:40:38.437 09:16:28 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:40:38.437 ************************************ 00:40:38.437 START TEST spdk_target_abort 00:40:38.437 ************************************ 00:40:38.437 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@1125 -- # spdk_target 00:40:38.437 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@43 -- # local name=spdk_target 00:40:38.438 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@45 -- # rpc_cmd bdev_nvme_attach_controller -t pcie -a 0000:65:00.0 -b spdk_target 00:40:38.438 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:38.438 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:40:39.010 spdk_targetn1 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@47 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:40:39.010 [2024-11-06 09:16:28.848812] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:testnqn -a -s SPDKISFASTANDAWESOME 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@49 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:testnqn spdk_targetn1 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@50 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:testnqn -t tcp -a 10.0.0.2 -s 4420 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:40:39.010 [2024-11-06 09:16:28.897122] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@52 -- # rabort tcp IPv4 10.0.0.2 4420 nqn.2016-06.io.spdk:testnqn 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.2 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@24 -- # local target r 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2' 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:40:39.010 09:16:28 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:40:39.274 [2024-11-06 09:16:29.136227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:189 nsid:1 lba:776 len:8 PRP1 0x200004abe000 PRP2 0x0 00:40:39.274 [2024-11-06 09:16:29.136255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:189 cdw0:0 sqhd:0062 p:1 m:0 dnr:0 00:40:39.274 [2024-11-06 09:16:29.138142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:190 nsid:1 lba:888 len:8 PRP1 0x200004ac6000 PRP2 0x0 00:40:39.274 [2024-11-06 09:16:29.138156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:190 cdw0:0 sqhd:0071 p:1 m:0 dnr:0 00:40:39.274 [2024-11-06 09:16:29.152197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:188 nsid:1 lba:1328 len:8 PRP1 0x200004ac0000 PRP2 0x0 00:40:39.274 [2024-11-06 09:16:29.152212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:188 cdw0:0 sqhd:00a8 p:1 m:0 dnr:0 00:40:39.274 [2024-11-06 09:16:29.160196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:188 nsid:1 lba:1584 len:8 PRP1 0x200004ac0000 PRP2 0x0 00:40:39.274 [2024-11-06 09:16:29.160210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:188 cdw0:0 sqhd:00c9 p:1 m:0 dnr:0 00:40:39.274 [2024-11-06 09:16:29.176194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:189 nsid:1 lba:2096 len:8 PRP1 0x200004ac2000 PRP2 0x0 00:40:39.274 [2024-11-06 09:16:29.176213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:189 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:40:39.274 [2024-11-06 09:16:29.200211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:189 nsid:1 lba:2944 len:8 PRP1 0x200004abe000 PRP2 0x0 00:40:39.274 [2024-11-06 09:16:29.200227] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:189 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:40:42.749 Initializing NVMe Controllers 00:40:42.749 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:40:42.749 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:40:42.749 Initialization complete. Launching workers. 00:40:42.749 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 13808, failed: 6 00:40:42.749 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 2758, failed to submit 11056 00:40:42.749 success 795, unsuccessful 1963, failed 0 00:40:42.749 09:16:32 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:40:42.749 09:16:32 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:40:42.749 [2024-11-06 09:16:32.387000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:181 nsid:1 lba:488 len:8 PRP1 0x200004e50000 PRP2 0x0 00:40:42.749 [2024-11-06 09:16:32.387035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:181 cdw0:0 sqhd:0048 p:1 m:0 dnr:0 00:40:42.749 [2024-11-06 09:16:32.403037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:176 nsid:1 lba:896 len:8 PRP1 0x200004e4e000 PRP2 0x0 00:40:42.749 [2024-11-06 09:16:32.403062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:176 cdw0:0 sqhd:0072 p:1 m:0 dnr:0 00:40:42.749 [2024-11-06 09:16:32.435011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:190 nsid:1 lba:1608 len:8 PRP1 0x200004e4a000 PRP2 0x0 00:40:42.750 [2024-11-06 09:16:32.435034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:190 cdw0:0 sqhd:00d2 p:1 m:0 dnr:0 00:40:42.750 [2024-11-06 09:16:32.521931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:188 nsid:1 lba:3600 len:8 PRP1 0x200004e4a000 PRP2 0x0 00:40:42.750 [2024-11-06 09:16:32.521955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:188 cdw0:0 sqhd:00cc p:0 m:0 dnr:0 00:40:45.295 [2024-11-06 09:16:34.938936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:4 cid:181 nsid:1 lba:57632 len:8 PRP1 0x200004e4a000 PRP2 0x0 00:40:45.295 [2024-11-06 09:16:34.938969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:4 cid:181 cdw0:0 sqhd:0026 p:1 m:0 dnr:0 00:40:45.556 Initializing NVMe Controllers 00:40:45.556 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:40:45.556 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:40:45.556 Initialization complete. Launching workers. 00:40:45.556 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 8420, failed: 5 00:40:45.556 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 1235, failed to submit 7190 00:40:45.556 success 346, unsuccessful 889, failed 0 00:40:45.556 09:16:35 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:40:45.556 09:16:35 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:40:45.817 [2024-11-06 09:16:35.704598] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:2 cid:154 nsid:1 lba:688 len:8 PRP1 0x200004af8000 PRP2 0x0 00:40:45.817 [2024-11-06 09:16:35.704624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - BY REQUEST (00/07) qid:2 cid:154 cdw0:0 sqhd:0058 p:1 m:0 dnr:0 00:40:49.119 Initializing NVMe Controllers 00:40:49.119 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:40:49.119 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:40:49.119 Initialization complete. Launching workers. 00:40:49.119 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 41799, failed: 1 00:40:49.119 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 2978, failed to submit 38822 00:40:49.119 success 588, unsuccessful 2390, failed 0 00:40:49.119 09:16:38 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@54 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:testnqn 00:40:49.119 09:16:38 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:49.119 09:16:38 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:40:49.119 09:16:38 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:49.119 09:16:38 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@55 -- # rpc_cmd bdev_nvme_detach_controller spdk_target 00:40:49.119 09:16:38 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@561 -- # xtrace_disable 00:40:49.119 09:16:38 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:40:50.508 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:40:50.508 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@61 -- # killprocess 2211875 00:40:50.508 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@950 -- # '[' -z 2211875 ']' 00:40:50.508 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@954 -- # kill -0 2211875 00:40:50.508 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@955 -- # uname 00:40:50.508 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:40:50.508 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2211875 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2211875' 00:40:50.768 killing process with pid 2211875 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@969 -- # kill 2211875 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@974 -- # wait 2211875 00:40:50.768 00:40:50.768 real 0m12.231s 00:40:50.768 user 0m49.963s 00:40:50.768 sys 0m1.826s 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@1126 -- # xtrace_disable 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:40:50.768 ************************************ 00:40:50.768 END TEST spdk_target_abort 00:40:50.768 ************************************ 00:40:50.768 09:16:40 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@81 -- # run_test kernel_target_abort kernel_target 00:40:50.768 09:16:40 nvmf_abort_qd_sizes -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:40:50.768 09:16:40 nvmf_abort_qd_sizes -- common/autotest_common.sh@1107 -- # xtrace_disable 00:40:50.768 09:16:40 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:40:50.768 ************************************ 00:40:50.768 START TEST kernel_target_abort 00:40:50.768 ************************************ 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1125 -- # kernel_target 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@65 -- # get_main_ns_ip 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@767 -- # local ip 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@768 -- # ip_candidates=() 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@768 -- # local -A ip_candidates 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@770 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@771 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@773 -- # [[ -z tcp ]] 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@773 -- # [[ -z NVMF_INITIATOR_IP ]] 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@774 -- # ip=NVMF_INITIATOR_IP 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@776 -- # [[ -z 10.0.0.1 ]] 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@781 -- # echo 10.0.0.1 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@65 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:40:50.768 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@658 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:40:50.769 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@660 -- # nvmet=/sys/kernel/config/nvmet 00:40:50.769 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@661 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:40:50.769 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@662 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:40:50.769 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@663 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:40:50.769 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@665 -- # local block nvme 00:40:50.769 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@667 -- # [[ ! -e /sys/module/nvmet ]] 00:40:50.769 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@668 -- # modprobe nvmet 00:40:51.029 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@671 -- # [[ -e /sys/kernel/config/nvmet ]] 00:40:51.029 09:16:40 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@673 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:40:54.332 Waiting for block devices as requested 00:40:54.332 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:40:54.332 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:40:54.332 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:40:54.332 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:40:54.332 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:40:54.332 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:40:54.593 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:40:54.593 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:40:54.593 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:40:54.854 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:40:54.854 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:40:55.115 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:40:55.115 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:40:55.115 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:40:55.115 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:40:55.377 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:40:55.377 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:40:55.637 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@676 -- # for block in /sys/block/nvme* 00:40:55.637 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@677 -- # [[ -e /sys/block/nvme0n1 ]] 00:40:55.637 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@678 -- # is_block_zoned nvme0n1 00:40:55.637 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1646 -- # local device=nvme0n1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1648 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1649 -- # [[ none != none ]] 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@679 -- # block_in_use nvme0n1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@381 -- # local block=nvme0n1 pt 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@390 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:40:55.638 No valid GPT data, bailing 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@394 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@394 -- # pt= 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@395 -- # return 1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@679 -- # nvme=/dev/nvme0n1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@682 -- # [[ -b /dev/nvme0n1 ]] 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@684 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@685 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@686 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@691 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@693 -- # echo 1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@694 -- # echo /dev/nvme0n1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@695 -- # echo 1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@697 -- # echo 10.0.0.1 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@698 -- # echo tcp 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@699 -- # echo 4420 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@700 -- # echo ipv4 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@703 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:40:55.638 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@706 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be --hostid=00d0226a-fbea-ec11-9bc7-a4bf019282be -a 10.0.0.1 -t tcp -s 4420 00:40:55.899 00:40:55.899 Discovery Log Number of Records 2, Generation counter 2 00:40:55.899 =====Discovery Log Entry 0====== 00:40:55.899 trtype: tcp 00:40:55.899 adrfam: ipv4 00:40:55.899 subtype: current discovery subsystem 00:40:55.899 treq: not specified, sq flow control disable supported 00:40:55.899 portid: 1 00:40:55.899 trsvcid: 4420 00:40:55.899 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:40:55.899 traddr: 10.0.0.1 00:40:55.899 eflags: none 00:40:55.899 sectype: none 00:40:55.899 =====Discovery Log Entry 1====== 00:40:55.899 trtype: tcp 00:40:55.899 adrfam: ipv4 00:40:55.899 subtype: nvme subsystem 00:40:55.899 treq: not specified, sq flow control disable supported 00:40:55.899 portid: 1 00:40:55.899 trsvcid: 4420 00:40:55.899 subnqn: nqn.2016-06.io.spdk:testnqn 00:40:55.899 traddr: 10.0.0.1 00:40:55.899 eflags: none 00:40:55.899 sectype: none 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@66 -- # rabort tcp IPv4 10.0.0.1 4420 nqn.2016-06.io.spdk:testnqn 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.1 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@24 -- # local target r 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1' 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420' 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:40:55.899 09:16:45 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:40:59.243 Initializing NVMe Controllers 00:40:59.243 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:40:59.243 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:40:59.243 Initialization complete. Launching workers. 00:40:59.243 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 66783, failed: 0 00:40:59.243 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 66783, failed to submit 0 00:40:59.243 success 0, unsuccessful 66783, failed 0 00:40:59.243 09:16:48 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:40:59.243 09:16:48 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:41:02.544 Initializing NVMe Controllers 00:41:02.544 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:41:02.544 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:41:02.544 Initialization complete. Launching workers. 00:41:02.544 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 107672, failed: 0 00:41:02.544 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 27122, failed to submit 80550 00:41:02.544 success 0, unsuccessful 27122, failed 0 00:41:02.544 09:16:51 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:41:02.544 09:16:51 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:41:05.087 Initializing NVMe Controllers 00:41:05.087 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:41:05.087 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:41:05.087 Initialization complete. Launching workers. 00:41:05.087 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 101427, failed: 0 00:41:05.087 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 25358, failed to submit 76069 00:41:05.087 success 0, unsuccessful 25358, failed 0 00:41:05.087 09:16:55 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@67 -- # clean_kernel_target 00:41:05.087 09:16:55 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@710 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:41:05.087 09:16:55 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@712 -- # echo 0 00:41:05.087 09:16:55 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@714 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:41:05.087 09:16:55 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@715 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:41:05.087 09:16:55 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@716 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:41:05.087 09:16:55 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@717 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:41:05.087 09:16:55 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@719 -- # modules=(/sys/module/nvmet/holders/*) 00:41:05.087 09:16:55 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@721 -- # modprobe -r nvmet_tcp nvmet 00:41:05.087 09:16:55 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@724 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:41:08.391 0000:80:01.6 (8086 0b00): ioatdma -> vfio-pci 00:41:08.391 0000:80:01.7 (8086 0b00): ioatdma -> vfio-pci 00:41:08.391 0000:80:01.4 (8086 0b00): ioatdma -> vfio-pci 00:41:08.391 0000:80:01.5 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:80:01.2 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:80:01.3 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:80:01.0 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:80:01.1 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:00:01.6 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:00:01.7 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:00:01.4 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:00:01.5 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:00:01.2 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:00:01.3 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:00:01.0 (8086 0b00): ioatdma -> vfio-pci 00:41:08.651 0000:00:01.1 (8086 0b00): ioatdma -> vfio-pci 00:41:10.562 0000:65:00.0 (144d a80a): nvme -> vfio-pci 00:41:10.822 00:41:10.822 real 0m19.888s 00:41:10.822 user 0m9.753s 00:41:10.822 sys 0m5.894s 00:41:10.822 09:17:00 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1126 -- # xtrace_disable 00:41:10.822 09:17:00 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@10 -- # set +x 00:41:10.822 ************************************ 00:41:10.822 END TEST kernel_target_abort 00:41:10.822 ************************************ 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@84 -- # nvmftestfini 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@514 -- # nvmfcleanup 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@121 -- # sync 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@123 -- # '[' tcp == tcp ']' 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@124 -- # set +e 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@125 -- # for i in {1..20} 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@126 -- # modprobe -v -r nvme-tcp 00:41:10.823 rmmod nvme_tcp 00:41:10.823 rmmod nvme_fabrics 00:41:10.823 rmmod nvme_keyring 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@127 -- # modprobe -v -r nvme-fabrics 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@128 -- # set -e 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@129 -- # return 0 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@515 -- # '[' -n 2211875 ']' 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@516 -- # killprocess 2211875 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- common/autotest_common.sh@950 -- # '[' -z 2211875 ']' 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- common/autotest_common.sh@954 -- # kill -0 2211875 00:41:10.823 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 954: kill: (2211875) - No such process 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- common/autotest_common.sh@977 -- # echo 'Process with pid 2211875 is not found' 00:41:10.823 Process with pid 2211875 is not found 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@518 -- # '[' iso == iso ']' 00:41:10.823 09:17:00 nvmf_abort_qd_sizes -- nvmf/common.sh@519 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:41:14.123 Waiting for block devices as requested 00:41:14.384 0000:80:01.6 (8086 0b00): vfio-pci -> ioatdma 00:41:14.384 0000:80:01.7 (8086 0b00): vfio-pci -> ioatdma 00:41:14.384 0000:80:01.4 (8086 0b00): vfio-pci -> ioatdma 00:41:14.384 0000:80:01.5 (8086 0b00): vfio-pci -> ioatdma 00:41:14.644 0000:80:01.2 (8086 0b00): vfio-pci -> ioatdma 00:41:14.644 0000:80:01.3 (8086 0b00): vfio-pci -> ioatdma 00:41:14.644 0000:80:01.0 (8086 0b00): vfio-pci -> ioatdma 00:41:14.904 0000:80:01.1 (8086 0b00): vfio-pci -> ioatdma 00:41:14.904 0000:65:00.0 (144d a80a): vfio-pci -> nvme 00:41:15.164 0000:00:01.6 (8086 0b00): vfio-pci -> ioatdma 00:41:15.164 0000:00:01.7 (8086 0b00): vfio-pci -> ioatdma 00:41:15.164 0000:00:01.4 (8086 0b00): vfio-pci -> ioatdma 00:41:15.424 0000:00:01.5 (8086 0b00): vfio-pci -> ioatdma 00:41:15.424 0000:00:01.2 (8086 0b00): vfio-pci -> ioatdma 00:41:15.424 0000:00:01.3 (8086 0b00): vfio-pci -> ioatdma 00:41:15.424 0000:00:01.0 (8086 0b00): vfio-pci -> ioatdma 00:41:15.684 0000:00:01.1 (8086 0b00): vfio-pci -> ioatdma 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- nvmf/common.sh@521 -- # [[ tcp == \t\c\p ]] 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- nvmf/common.sh@522 -- # nvmf_tcp_fini 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- nvmf/common.sh@297 -- # iptr 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- nvmf/common.sh@789 -- # iptables-save 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- nvmf/common.sh@789 -- # grep -v SPDK_NVMF 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- nvmf/common.sh@789 -- # iptables-restore 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- nvmf/common.sh@298 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s\_\s\p\d\k ]] 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- nvmf/common.sh@302 -- # remove_spdk_ns 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- nvmf/common.sh@654 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:41:15.946 09:17:05 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:41:17.859 09:17:07 nvmf_abort_qd_sizes -- nvmf/common.sh@303 -- # ip -4 addr flush cvl_0_1 00:41:17.859 00:41:17.859 real 0m51.607s 00:41:17.859 user 1m5.079s 00:41:17.859 sys 0m18.516s 00:41:17.859 09:17:07 nvmf_abort_qd_sizes -- common/autotest_common.sh@1126 -- # xtrace_disable 00:41:17.859 09:17:07 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:41:17.859 ************************************ 00:41:17.859 END TEST nvmf_abort_qd_sizes 00:41:17.859 ************************************ 00:41:18.119 09:17:08 -- spdk/autotest.sh@288 -- # run_test keyring_file /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:41:18.119 09:17:08 -- common/autotest_common.sh@1101 -- # '[' 2 -le 1 ']' 00:41:18.119 09:17:08 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:41:18.119 09:17:08 -- common/autotest_common.sh@10 -- # set +x 00:41:18.119 ************************************ 00:41:18.119 START TEST keyring_file 00:41:18.119 ************************************ 00:41:18.119 09:17:08 keyring_file -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:41:18.119 * Looking for test storage... 00:41:18.119 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:41:18.119 09:17:08 keyring_file -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:41:18.119 09:17:08 keyring_file -- common/autotest_common.sh@1689 -- # lcov --version 00:41:18.119 09:17:08 keyring_file -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:41:18.119 09:17:08 keyring_file -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:41:18.119 09:17:08 keyring_file -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:41:18.119 09:17:08 keyring_file -- scripts/common.sh@333 -- # local ver1 ver1_l 00:41:18.119 09:17:08 keyring_file -- scripts/common.sh@334 -- # local ver2 ver2_l 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@336 -- # IFS=.-: 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@336 -- # read -ra ver1 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@337 -- # IFS=.-: 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@337 -- # read -ra ver2 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@338 -- # local 'op=<' 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@340 -- # ver1_l=2 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@341 -- # ver2_l=1 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@344 -- # case "$op" in 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@345 -- # : 1 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@364 -- # (( v = 0 )) 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@365 -- # decimal 1 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@353 -- # local d=1 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:41:18.120 09:17:08 keyring_file -- scripts/common.sh@355 -- # echo 1 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@365 -- # ver1[v]=1 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@366 -- # decimal 2 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@353 -- # local d=2 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@355 -- # echo 2 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@366 -- # ver2[v]=2 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@368 -- # return 0 00:41:18.381 09:17:08 keyring_file -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:41:18.381 09:17:08 keyring_file -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:41:18.381 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:18.381 --rc genhtml_branch_coverage=1 00:41:18.381 --rc genhtml_function_coverage=1 00:41:18.381 --rc genhtml_legend=1 00:41:18.381 --rc geninfo_all_blocks=1 00:41:18.381 --rc geninfo_unexecuted_blocks=1 00:41:18.381 00:41:18.381 ' 00:41:18.381 09:17:08 keyring_file -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:41:18.381 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:18.381 --rc genhtml_branch_coverage=1 00:41:18.381 --rc genhtml_function_coverage=1 00:41:18.381 --rc genhtml_legend=1 00:41:18.381 --rc geninfo_all_blocks=1 00:41:18.381 --rc geninfo_unexecuted_blocks=1 00:41:18.381 00:41:18.381 ' 00:41:18.381 09:17:08 keyring_file -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:41:18.381 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:18.381 --rc genhtml_branch_coverage=1 00:41:18.381 --rc genhtml_function_coverage=1 00:41:18.381 --rc genhtml_legend=1 00:41:18.381 --rc geninfo_all_blocks=1 00:41:18.381 --rc geninfo_unexecuted_blocks=1 00:41:18.381 00:41:18.381 ' 00:41:18.381 09:17:08 keyring_file -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:41:18.381 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:18.381 --rc genhtml_branch_coverage=1 00:41:18.381 --rc genhtml_function_coverage=1 00:41:18.381 --rc genhtml_legend=1 00:41:18.381 --rc geninfo_all_blocks=1 00:41:18.381 --rc geninfo_unexecuted_blocks=1 00:41:18.381 00:41:18.381 ' 00:41:18.381 09:17:08 keyring_file -- keyring/file.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@7 -- # uname -s 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@15 -- # shopt -s extglob 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:41:18.381 09:17:08 keyring_file -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:41:18.381 09:17:08 keyring_file -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:18.381 09:17:08 keyring_file -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:18.381 09:17:08 keyring_file -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:18.381 09:17:08 keyring_file -- paths/export.sh@5 -- # export PATH 00:41:18.381 09:17:08 keyring_file -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@51 -- # : 0 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:41:18.381 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@55 -- # have_pci_nics=0 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:41:18.381 09:17:08 keyring_file -- keyring/file.sh@13 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:41:18.381 09:17:08 keyring_file -- keyring/file.sh@14 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:41:18.381 09:17:08 keyring_file -- keyring/file.sh@15 -- # key0=00112233445566778899aabbccddeeff 00:41:18.381 09:17:08 keyring_file -- keyring/file.sh@16 -- # key1=112233445566778899aabbccddeeff00 00:41:18.381 09:17:08 keyring_file -- keyring/file.sh@24 -- # trap cleanup EXIT 00:41:18.381 09:17:08 keyring_file -- keyring/file.sh@26 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@17 -- # name=key0 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@17 -- # digest=0 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@18 -- # mktemp 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.0xR7ohLevf 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@728 -- # local prefix key digest 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@730 -- # key=00112233445566778899aabbccddeeff 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@730 -- # digest=0 00:41:18.381 09:17:08 keyring_file -- nvmf/common.sh@731 -- # python - 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.0xR7ohLevf 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.0xR7ohLevf 00:41:18.381 09:17:08 keyring_file -- keyring/file.sh@26 -- # key0path=/tmp/tmp.0xR7ohLevf 00:41:18.381 09:17:08 keyring_file -- keyring/file.sh@27 -- # prep_key key1 112233445566778899aabbccddeeff00 0 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@17 -- # name=key1 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@17 -- # digest=0 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@18 -- # mktemp 00:41:18.381 09:17:08 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.wQtzL8FpXn 00:41:18.382 09:17:08 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:41:18.382 09:17:08 keyring_file -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:41:18.382 09:17:08 keyring_file -- nvmf/common.sh@728 -- # local prefix key digest 00:41:18.382 09:17:08 keyring_file -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:41:18.382 09:17:08 keyring_file -- nvmf/common.sh@730 -- # key=112233445566778899aabbccddeeff00 00:41:18.382 09:17:08 keyring_file -- nvmf/common.sh@730 -- # digest=0 00:41:18.382 09:17:08 keyring_file -- nvmf/common.sh@731 -- # python - 00:41:18.382 09:17:08 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.wQtzL8FpXn 00:41:18.382 09:17:08 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.wQtzL8FpXn 00:41:18.382 09:17:08 keyring_file -- keyring/file.sh@27 -- # key1path=/tmp/tmp.wQtzL8FpXn 00:41:18.382 09:17:08 keyring_file -- keyring/file.sh@30 -- # tgtpid=2222274 00:41:18.382 09:17:08 keyring_file -- keyring/file.sh@32 -- # waitforlisten 2222274 00:41:18.382 09:17:08 keyring_file -- keyring/file.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:41:18.382 09:17:08 keyring_file -- common/autotest_common.sh@831 -- # '[' -z 2222274 ']' 00:41:18.382 09:17:08 keyring_file -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:41:18.382 09:17:08 keyring_file -- common/autotest_common.sh@836 -- # local max_retries=100 00:41:18.382 09:17:08 keyring_file -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:41:18.382 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:41:18.382 09:17:08 keyring_file -- common/autotest_common.sh@840 -- # xtrace_disable 00:41:18.382 09:17:08 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:41:18.382 [2024-11-06 09:17:08.456909] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:41:18.382 [2024-11-06 09:17:08.456966] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2222274 ] 00:41:18.642 [2024-11-06 09:17:08.527919] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:18.642 [2024-11-06 09:17:08.563850] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@864 -- # return 0 00:41:19.212 09:17:09 keyring_file -- keyring/file.sh@33 -- # rpc_cmd 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@561 -- # xtrace_disable 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:41:19.212 [2024-11-06 09:17:09.248449] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:41:19.212 null0 00:41:19.212 [2024-11-06 09:17:09.280503] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:41:19.212 [2024-11-06 09:17:09.280875] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:41:19.212 09:17:09 keyring_file -- keyring/file.sh@44 -- # NOT rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@650 -- # local es=0 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@652 -- # valid_exec_arg rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@638 -- # local arg=rpc_cmd 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@642 -- # type -t rpc_cmd 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@653 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:41:19.212 09:17:09 keyring_file -- common/autotest_common.sh@561 -- # xtrace_disable 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:41:19.213 [2024-11-06 09:17:09.312572] nvmf_rpc.c: 762:nvmf_rpc_listen_paused: *ERROR*: Listener already exists 00:41:19.213 request: 00:41:19.213 { 00:41:19.213 "nqn": "nqn.2016-06.io.spdk:cnode0", 00:41:19.213 "secure_channel": false, 00:41:19.213 "listen_address": { 00:41:19.213 "trtype": "tcp", 00:41:19.213 "traddr": "127.0.0.1", 00:41:19.213 "trsvcid": "4420" 00:41:19.213 }, 00:41:19.213 "method": "nvmf_subsystem_add_listener", 00:41:19.213 "req_id": 1 00:41:19.213 } 00:41:19.213 Got JSON-RPC error response 00:41:19.213 response: 00:41:19.213 { 00:41:19.213 "code": -32602, 00:41:19.213 "message": "Invalid parameters" 00:41:19.213 } 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@589 -- # [[ 1 == 0 ]] 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@653 -- # es=1 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:41:19.213 09:17:09 keyring_file -- keyring/file.sh@47 -- # bperfpid=2222292 00:41:19.213 09:17:09 keyring_file -- keyring/file.sh@49 -- # waitforlisten 2222292 /var/tmp/bperf.sock 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@831 -- # '[' -z 2222292 ']' 00:41:19.213 09:17:09 keyring_file -- keyring/file.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@836 -- # local max_retries=100 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:41:19.213 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@840 -- # xtrace_disable 00:41:19.213 09:17:09 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:41:19.473 [2024-11-06 09:17:09.371422] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:41:19.473 [2024-11-06 09:17:09.371472] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2222292 ] 00:41:19.473 [2024-11-06 09:17:09.458920] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:19.473 [2024-11-06 09:17:09.495065] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:41:20.043 09:17:10 keyring_file -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:41:20.043 09:17:10 keyring_file -- common/autotest_common.sh@864 -- # return 0 00:41:20.043 09:17:10 keyring_file -- keyring/file.sh@50 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.0xR7ohLevf 00:41:20.043 09:17:10 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.0xR7ohLevf 00:41:20.303 09:17:10 keyring_file -- keyring/file.sh@51 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.wQtzL8FpXn 00:41:20.303 09:17:10 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.wQtzL8FpXn 00:41:20.564 09:17:10 keyring_file -- keyring/file.sh@52 -- # get_key key0 00:41:20.564 09:17:10 keyring_file -- keyring/file.sh@52 -- # jq -r .path 00:41:20.564 09:17:10 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:20.564 09:17:10 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:41:20.564 09:17:10 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:20.564 09:17:10 keyring_file -- keyring/file.sh@52 -- # [[ /tmp/tmp.0xR7ohLevf == \/\t\m\p\/\t\m\p\.\0\x\R\7\o\h\L\e\v\f ]] 00:41:20.564 09:17:10 keyring_file -- keyring/file.sh@53 -- # get_key key1 00:41:20.564 09:17:10 keyring_file -- keyring/file.sh@53 -- # jq -r .path 00:41:20.564 09:17:10 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:20.564 09:17:10 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:41:20.564 09:17:10 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:20.824 09:17:10 keyring_file -- keyring/file.sh@53 -- # [[ /tmp/tmp.wQtzL8FpXn == \/\t\m\p\/\t\m\p\.\w\Q\t\z\L\8\F\p\X\n ]] 00:41:20.824 09:17:10 keyring_file -- keyring/file.sh@54 -- # get_refcnt key0 00:41:20.824 09:17:10 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:41:20.824 09:17:10 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:20.824 09:17:10 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:41:20.824 09:17:10 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:20.824 09:17:10 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:21.084 09:17:11 keyring_file -- keyring/file.sh@54 -- # (( 1 == 1 )) 00:41:21.084 09:17:11 keyring_file -- keyring/file.sh@55 -- # get_refcnt key1 00:41:21.084 09:17:11 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:41:21.084 09:17:11 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:21.084 09:17:11 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:21.084 09:17:11 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:41:21.084 09:17:11 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:21.084 09:17:11 keyring_file -- keyring/file.sh@55 -- # (( 1 == 1 )) 00:41:21.084 09:17:11 keyring_file -- keyring/file.sh@58 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:41:21.084 09:17:11 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:41:21.344 [2024-11-06 09:17:11.330452] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:41:21.344 nvme0n1 00:41:21.344 09:17:11 keyring_file -- keyring/file.sh@60 -- # get_refcnt key0 00:41:21.344 09:17:11 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:41:21.344 09:17:11 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:21.344 09:17:11 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:21.344 09:17:11 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:21.344 09:17:11 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:41:21.604 09:17:11 keyring_file -- keyring/file.sh@60 -- # (( 2 == 2 )) 00:41:21.604 09:17:11 keyring_file -- keyring/file.sh@61 -- # get_refcnt key1 00:41:21.604 09:17:11 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:41:21.604 09:17:11 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:41:21.604 09:17:11 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:21.604 09:17:11 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:21.604 09:17:11 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:21.867 09:17:11 keyring_file -- keyring/file.sh@61 -- # (( 1 == 1 )) 00:41:21.867 09:17:11 keyring_file -- keyring/file.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:41:21.867 Running I/O for 1 seconds... 00:41:22.806 15672.00 IOPS, 61.22 MiB/s 00:41:22.807 Latency(us) 00:41:22.807 [2024-11-06T08:17:12.920Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:22.807 Job: nvme0n1 (Core Mask 0x2, workload: randrw, percentage: 50, depth: 128, IO size: 4096) 00:41:22.807 nvme0n1 : 1.01 15687.75 61.28 0.00 0.00 8127.28 5761.71 18350.08 00:41:22.807 [2024-11-06T08:17:12.920Z] =================================================================================================================== 00:41:22.807 [2024-11-06T08:17:12.920Z] Total : 15687.75 61.28 0.00 0.00 8127.28 5761.71 18350.08 00:41:22.807 { 00:41:22.807 "results": [ 00:41:22.807 { 00:41:22.807 "job": "nvme0n1", 00:41:22.807 "core_mask": "0x2", 00:41:22.807 "workload": "randrw", 00:41:22.807 "percentage": 50, 00:41:22.807 "status": "finished", 00:41:22.807 "queue_depth": 128, 00:41:22.807 "io_size": 4096, 00:41:22.807 "runtime": 1.007283, 00:41:22.807 "iops": 15687.746144827222, 00:41:22.807 "mibps": 61.28025837823134, 00:41:22.807 "io_failed": 0, 00:41:22.807 "io_timeout": 0, 00:41:22.807 "avg_latency_us": 8127.280158629708, 00:41:22.807 "min_latency_us": 5761.706666666667, 00:41:22.807 "max_latency_us": 18350.08 00:41:22.807 } 00:41:22.807 ], 00:41:22.807 "core_count": 1 00:41:22.807 } 00:41:22.807 09:17:12 keyring_file -- keyring/file.sh@65 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:41:22.807 09:17:12 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:41:23.067 09:17:13 keyring_file -- keyring/file.sh@66 -- # get_refcnt key0 00:41:23.067 09:17:13 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:41:23.067 09:17:13 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:23.067 09:17:13 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:23.067 09:17:13 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:41:23.067 09:17:13 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:23.328 09:17:13 keyring_file -- keyring/file.sh@66 -- # (( 1 == 1 )) 00:41:23.328 09:17:13 keyring_file -- keyring/file.sh@67 -- # get_refcnt key1 00:41:23.328 09:17:13 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:41:23.328 09:17:13 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:23.328 09:17:13 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:23.328 09:17:13 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:41:23.328 09:17:13 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:23.328 09:17:13 keyring_file -- keyring/file.sh@67 -- # (( 1 == 1 )) 00:41:23.328 09:17:13 keyring_file -- keyring/file.sh@70 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:41:23.328 09:17:13 keyring_file -- common/autotest_common.sh@650 -- # local es=0 00:41:23.328 09:17:13 keyring_file -- common/autotest_common.sh@652 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:41:23.328 09:17:13 keyring_file -- common/autotest_common.sh@638 -- # local arg=bperf_cmd 00:41:23.328 09:17:13 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:41:23.328 09:17:13 keyring_file -- common/autotest_common.sh@642 -- # type -t bperf_cmd 00:41:23.328 09:17:13 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:41:23.328 09:17:13 keyring_file -- common/autotest_common.sh@653 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:41:23.328 09:17:13 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:41:23.588 [2024-11-06 09:17:13.565616] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:41:23.588 [2024-11-06 09:17:13.566332] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x20796f0 (107): Transport endpoint is not connected 00:41:23.588 [2024-11-06 09:17:13.567327] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x20796f0 (9): Bad file descriptor 00:41:23.588 [2024-11-06 09:17:13.568329] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] Ctrlr is in error state 00:41:23.588 [2024-11-06 09:17:13.568336] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:41:23.588 [2024-11-06 09:17:13.568343] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, Operation not permitted 00:41:23.588 [2024-11-06 09:17:13.568349] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] in failed state. 00:41:23.588 request: 00:41:23.588 { 00:41:23.588 "name": "nvme0", 00:41:23.588 "trtype": "tcp", 00:41:23.588 "traddr": "127.0.0.1", 00:41:23.588 "adrfam": "ipv4", 00:41:23.588 "trsvcid": "4420", 00:41:23.588 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:41:23.588 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:41:23.588 "prchk_reftag": false, 00:41:23.588 "prchk_guard": false, 00:41:23.588 "hdgst": false, 00:41:23.588 "ddgst": false, 00:41:23.588 "psk": "key1", 00:41:23.588 "allow_unrecognized_csi": false, 00:41:23.588 "method": "bdev_nvme_attach_controller", 00:41:23.588 "req_id": 1 00:41:23.588 } 00:41:23.588 Got JSON-RPC error response 00:41:23.588 response: 00:41:23.588 { 00:41:23.588 "code": -5, 00:41:23.588 "message": "Input/output error" 00:41:23.588 } 00:41:23.588 09:17:13 keyring_file -- common/autotest_common.sh@653 -- # es=1 00:41:23.588 09:17:13 keyring_file -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:41:23.588 09:17:13 keyring_file -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:41:23.589 09:17:13 keyring_file -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:41:23.589 09:17:13 keyring_file -- keyring/file.sh@72 -- # get_refcnt key0 00:41:23.589 09:17:13 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:41:23.589 09:17:13 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:23.589 09:17:13 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:23.589 09:17:13 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:23.589 09:17:13 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:41:23.849 09:17:13 keyring_file -- keyring/file.sh@72 -- # (( 1 == 1 )) 00:41:23.849 09:17:13 keyring_file -- keyring/file.sh@73 -- # get_refcnt key1 00:41:23.849 09:17:13 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:41:23.849 09:17:13 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:23.849 09:17:13 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:23.849 09:17:13 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:23.849 09:17:13 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:41:23.849 09:17:13 keyring_file -- keyring/file.sh@73 -- # (( 1 == 1 )) 00:41:23.849 09:17:13 keyring_file -- keyring/file.sh@76 -- # bperf_cmd keyring_file_remove_key key0 00:41:23.849 09:17:13 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:41:24.109 09:17:14 keyring_file -- keyring/file.sh@77 -- # bperf_cmd keyring_file_remove_key key1 00:41:24.109 09:17:14 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key1 00:41:24.369 09:17:14 keyring_file -- keyring/file.sh@78 -- # bperf_cmd keyring_get_keys 00:41:24.369 09:17:14 keyring_file -- keyring/file.sh@78 -- # jq length 00:41:24.369 09:17:14 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:24.369 09:17:14 keyring_file -- keyring/file.sh@78 -- # (( 0 == 0 )) 00:41:24.369 09:17:14 keyring_file -- keyring/file.sh@81 -- # chmod 0660 /tmp/tmp.0xR7ohLevf 00:41:24.369 09:17:14 keyring_file -- keyring/file.sh@82 -- # NOT bperf_cmd keyring_file_add_key key0 /tmp/tmp.0xR7ohLevf 00:41:24.369 09:17:14 keyring_file -- common/autotest_common.sh@650 -- # local es=0 00:41:24.369 09:17:14 keyring_file -- common/autotest_common.sh@652 -- # valid_exec_arg bperf_cmd keyring_file_add_key key0 /tmp/tmp.0xR7ohLevf 00:41:24.369 09:17:14 keyring_file -- common/autotest_common.sh@638 -- # local arg=bperf_cmd 00:41:24.369 09:17:14 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:41:24.369 09:17:14 keyring_file -- common/autotest_common.sh@642 -- # type -t bperf_cmd 00:41:24.369 09:17:14 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:41:24.369 09:17:14 keyring_file -- common/autotest_common.sh@653 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.0xR7ohLevf 00:41:24.369 09:17:14 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.0xR7ohLevf 00:41:24.629 [2024-11-06 09:17:14.584937] keyring.c: 36:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.0xR7ohLevf': 0100660 00:41:24.629 [2024-11-06 09:17:14.584954] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:41:24.629 request: 00:41:24.629 { 00:41:24.629 "name": "key0", 00:41:24.629 "path": "/tmp/tmp.0xR7ohLevf", 00:41:24.629 "method": "keyring_file_add_key", 00:41:24.629 "req_id": 1 00:41:24.629 } 00:41:24.629 Got JSON-RPC error response 00:41:24.629 response: 00:41:24.629 { 00:41:24.629 "code": -1, 00:41:24.629 "message": "Operation not permitted" 00:41:24.629 } 00:41:24.629 09:17:14 keyring_file -- common/autotest_common.sh@653 -- # es=1 00:41:24.629 09:17:14 keyring_file -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:41:24.629 09:17:14 keyring_file -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:41:24.629 09:17:14 keyring_file -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:41:24.629 09:17:14 keyring_file -- keyring/file.sh@85 -- # chmod 0600 /tmp/tmp.0xR7ohLevf 00:41:24.629 09:17:14 keyring_file -- keyring/file.sh@86 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.0xR7ohLevf 00:41:24.629 09:17:14 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.0xR7ohLevf 00:41:24.889 09:17:14 keyring_file -- keyring/file.sh@87 -- # rm -f /tmp/tmp.0xR7ohLevf 00:41:24.889 09:17:14 keyring_file -- keyring/file.sh@89 -- # get_refcnt key0 00:41:24.889 09:17:14 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:41:24.889 09:17:14 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:24.889 09:17:14 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:24.889 09:17:14 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:41:24.889 09:17:14 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:24.889 09:17:14 keyring_file -- keyring/file.sh@89 -- # (( 1 == 1 )) 00:41:24.889 09:17:14 keyring_file -- keyring/file.sh@91 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:41:24.889 09:17:14 keyring_file -- common/autotest_common.sh@650 -- # local es=0 00:41:24.889 09:17:14 keyring_file -- common/autotest_common.sh@652 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:41:24.889 09:17:14 keyring_file -- common/autotest_common.sh@638 -- # local arg=bperf_cmd 00:41:24.889 09:17:14 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:41:24.889 09:17:14 keyring_file -- common/autotest_common.sh@642 -- # type -t bperf_cmd 00:41:24.889 09:17:14 keyring_file -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:41:24.889 09:17:14 keyring_file -- common/autotest_common.sh@653 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:41:24.889 09:17:14 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:41:25.150 [2024-11-06 09:17:15.110270] keyring.c: 31:keyring_file_check_path: *ERROR*: Could not stat key file '/tmp/tmp.0xR7ohLevf': No such file or directory 00:41:25.150 [2024-11-06 09:17:15.110283] nvme_tcp.c:2498:nvme_tcp_generate_tls_credentials: *ERROR*: Failed to obtain key 'key0': No such file or directory 00:41:25.150 [2024-11-06 09:17:15.110296] nvme.c: 682:nvme_ctrlr_probe: *ERROR*: Failed to construct NVMe controller for SSD: 127.0.0.1 00:41:25.150 [2024-11-06 09:17:15.110301] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, No such device 00:41:25.150 [2024-11-06 09:17:15.110311] nvme.c: 831:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:41:25.150 [2024-11-06 09:17:15.110317] bdev_nvme.c:6576:spdk_bdev_nvme_create: *ERROR*: No controller was found with provided trid (traddr: 127.0.0.1) 00:41:25.150 request: 00:41:25.150 { 00:41:25.150 "name": "nvme0", 00:41:25.150 "trtype": "tcp", 00:41:25.150 "traddr": "127.0.0.1", 00:41:25.150 "adrfam": "ipv4", 00:41:25.150 "trsvcid": "4420", 00:41:25.150 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:41:25.150 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:41:25.150 "prchk_reftag": false, 00:41:25.150 "prchk_guard": false, 00:41:25.150 "hdgst": false, 00:41:25.150 "ddgst": false, 00:41:25.150 "psk": "key0", 00:41:25.150 "allow_unrecognized_csi": false, 00:41:25.150 "method": "bdev_nvme_attach_controller", 00:41:25.150 "req_id": 1 00:41:25.150 } 00:41:25.150 Got JSON-RPC error response 00:41:25.150 response: 00:41:25.150 { 00:41:25.150 "code": -19, 00:41:25.150 "message": "No such device" 00:41:25.150 } 00:41:25.150 09:17:15 keyring_file -- common/autotest_common.sh@653 -- # es=1 00:41:25.150 09:17:15 keyring_file -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:41:25.150 09:17:15 keyring_file -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:41:25.150 09:17:15 keyring_file -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:41:25.150 09:17:15 keyring_file -- keyring/file.sh@93 -- # bperf_cmd keyring_file_remove_key key0 00:41:25.150 09:17:15 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:41:25.410 09:17:15 keyring_file -- keyring/file.sh@96 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:41:25.410 09:17:15 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:41:25.410 09:17:15 keyring_file -- keyring/common.sh@17 -- # name=key0 00:41:25.410 09:17:15 keyring_file -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:41:25.410 09:17:15 keyring_file -- keyring/common.sh@17 -- # digest=0 00:41:25.410 09:17:15 keyring_file -- keyring/common.sh@18 -- # mktemp 00:41:25.410 09:17:15 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.aEuhojhxs3 00:41:25.410 09:17:15 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:41:25.410 09:17:15 keyring_file -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:41:25.410 09:17:15 keyring_file -- nvmf/common.sh@728 -- # local prefix key digest 00:41:25.410 09:17:15 keyring_file -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:41:25.410 09:17:15 keyring_file -- nvmf/common.sh@730 -- # key=00112233445566778899aabbccddeeff 00:41:25.410 09:17:15 keyring_file -- nvmf/common.sh@730 -- # digest=0 00:41:25.410 09:17:15 keyring_file -- nvmf/common.sh@731 -- # python - 00:41:25.410 09:17:15 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.aEuhojhxs3 00:41:25.410 09:17:15 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.aEuhojhxs3 00:41:25.410 09:17:15 keyring_file -- keyring/file.sh@96 -- # key0path=/tmp/tmp.aEuhojhxs3 00:41:25.410 09:17:15 keyring_file -- keyring/file.sh@97 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.aEuhojhxs3 00:41:25.410 09:17:15 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.aEuhojhxs3 00:41:25.410 09:17:15 keyring_file -- keyring/file.sh@98 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:41:25.411 09:17:15 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:41:25.670 nvme0n1 00:41:25.670 09:17:15 keyring_file -- keyring/file.sh@100 -- # get_refcnt key0 00:41:25.670 09:17:15 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:41:25.670 09:17:15 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:25.670 09:17:15 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:25.670 09:17:15 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:25.670 09:17:15 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:41:25.930 09:17:15 keyring_file -- keyring/file.sh@100 -- # (( 2 == 2 )) 00:41:25.930 09:17:15 keyring_file -- keyring/file.sh@101 -- # bperf_cmd keyring_file_remove_key key0 00:41:25.930 09:17:15 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:41:26.190 09:17:16 keyring_file -- keyring/file.sh@102 -- # get_key key0 00:41:26.190 09:17:16 keyring_file -- keyring/file.sh@102 -- # jq -r .removed 00:41:26.190 09:17:16 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:26.190 09:17:16 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:26.190 09:17:16 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:41:26.190 09:17:16 keyring_file -- keyring/file.sh@102 -- # [[ true == \t\r\u\e ]] 00:41:26.190 09:17:16 keyring_file -- keyring/file.sh@103 -- # get_refcnt key0 00:41:26.190 09:17:16 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:41:26.190 09:17:16 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:26.190 09:17:16 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:26.190 09:17:16 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:41:26.190 09:17:16 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:26.451 09:17:16 keyring_file -- keyring/file.sh@103 -- # (( 1 == 1 )) 00:41:26.451 09:17:16 keyring_file -- keyring/file.sh@104 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:41:26.451 09:17:16 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:41:26.711 09:17:16 keyring_file -- keyring/file.sh@105 -- # bperf_cmd keyring_get_keys 00:41:26.711 09:17:16 keyring_file -- keyring/file.sh@105 -- # jq length 00:41:26.711 09:17:16 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:26.711 09:17:16 keyring_file -- keyring/file.sh@105 -- # (( 0 == 0 )) 00:41:26.711 09:17:16 keyring_file -- keyring/file.sh@108 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.aEuhojhxs3 00:41:26.711 09:17:16 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.aEuhojhxs3 00:41:26.971 09:17:16 keyring_file -- keyring/file.sh@109 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.wQtzL8FpXn 00:41:26.971 09:17:16 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.wQtzL8FpXn 00:41:27.231 09:17:17 keyring_file -- keyring/file.sh@110 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:41:27.231 09:17:17 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:41:27.231 nvme0n1 00:41:27.491 09:17:17 keyring_file -- keyring/file.sh@113 -- # bperf_cmd save_config 00:41:27.491 09:17:17 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock save_config 00:41:27.491 09:17:17 keyring_file -- keyring/file.sh@113 -- # config='{ 00:41:27.491 "subsystems": [ 00:41:27.491 { 00:41:27.491 "subsystem": "keyring", 00:41:27.491 "config": [ 00:41:27.491 { 00:41:27.491 "method": "keyring_file_add_key", 00:41:27.491 "params": { 00:41:27.491 "name": "key0", 00:41:27.491 "path": "/tmp/tmp.aEuhojhxs3" 00:41:27.491 } 00:41:27.491 }, 00:41:27.491 { 00:41:27.491 "method": "keyring_file_add_key", 00:41:27.491 "params": { 00:41:27.491 "name": "key1", 00:41:27.491 "path": "/tmp/tmp.wQtzL8FpXn" 00:41:27.491 } 00:41:27.491 } 00:41:27.491 ] 00:41:27.491 }, 00:41:27.491 { 00:41:27.491 "subsystem": "iobuf", 00:41:27.491 "config": [ 00:41:27.491 { 00:41:27.491 "method": "iobuf_set_options", 00:41:27.491 "params": { 00:41:27.491 "small_pool_count": 8192, 00:41:27.491 "large_pool_count": 1024, 00:41:27.491 "small_bufsize": 8192, 00:41:27.491 "large_bufsize": 135168, 00:41:27.491 "enable_numa": false 00:41:27.491 } 00:41:27.491 } 00:41:27.491 ] 00:41:27.491 }, 00:41:27.491 { 00:41:27.491 "subsystem": "sock", 00:41:27.491 "config": [ 00:41:27.491 { 00:41:27.492 "method": "sock_set_default_impl", 00:41:27.492 "params": { 00:41:27.492 "impl_name": "posix" 00:41:27.492 } 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "method": "sock_impl_set_options", 00:41:27.492 "params": { 00:41:27.492 "impl_name": "ssl", 00:41:27.492 "recv_buf_size": 4096, 00:41:27.492 "send_buf_size": 4096, 00:41:27.492 "enable_recv_pipe": true, 00:41:27.492 "enable_quickack": false, 00:41:27.492 "enable_placement_id": 0, 00:41:27.492 "enable_zerocopy_send_server": true, 00:41:27.492 "enable_zerocopy_send_client": false, 00:41:27.492 "zerocopy_threshold": 0, 00:41:27.492 "tls_version": 0, 00:41:27.492 "enable_ktls": false 00:41:27.492 } 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "method": "sock_impl_set_options", 00:41:27.492 "params": { 00:41:27.492 "impl_name": "posix", 00:41:27.492 "recv_buf_size": 2097152, 00:41:27.492 "send_buf_size": 2097152, 00:41:27.492 "enable_recv_pipe": true, 00:41:27.492 "enable_quickack": false, 00:41:27.492 "enable_placement_id": 0, 00:41:27.492 "enable_zerocopy_send_server": true, 00:41:27.492 "enable_zerocopy_send_client": false, 00:41:27.492 "zerocopy_threshold": 0, 00:41:27.492 "tls_version": 0, 00:41:27.492 "enable_ktls": false 00:41:27.492 } 00:41:27.492 } 00:41:27.492 ] 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "subsystem": "vmd", 00:41:27.492 "config": [] 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "subsystem": "accel", 00:41:27.492 "config": [ 00:41:27.492 { 00:41:27.492 "method": "accel_set_options", 00:41:27.492 "params": { 00:41:27.492 "small_cache_size": 128, 00:41:27.492 "large_cache_size": 16, 00:41:27.492 "task_count": 2048, 00:41:27.492 "sequence_count": 2048, 00:41:27.492 "buf_count": 2048 00:41:27.492 } 00:41:27.492 } 00:41:27.492 ] 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "subsystem": "bdev", 00:41:27.492 "config": [ 00:41:27.492 { 00:41:27.492 "method": "bdev_set_options", 00:41:27.492 "params": { 00:41:27.492 "bdev_io_pool_size": 65535, 00:41:27.492 "bdev_io_cache_size": 256, 00:41:27.492 "bdev_auto_examine": true, 00:41:27.492 "iobuf_small_cache_size": 128, 00:41:27.492 "iobuf_large_cache_size": 16 00:41:27.492 } 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "method": "bdev_raid_set_options", 00:41:27.492 "params": { 00:41:27.492 "process_window_size_kb": 1024, 00:41:27.492 "process_max_bandwidth_mb_sec": 0 00:41:27.492 } 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "method": "bdev_iscsi_set_options", 00:41:27.492 "params": { 00:41:27.492 "timeout_sec": 30 00:41:27.492 } 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "method": "bdev_nvme_set_options", 00:41:27.492 "params": { 00:41:27.492 "action_on_timeout": "none", 00:41:27.492 "timeout_us": 0, 00:41:27.492 "timeout_admin_us": 0, 00:41:27.492 "keep_alive_timeout_ms": 10000, 00:41:27.492 "arbitration_burst": 0, 00:41:27.492 "low_priority_weight": 0, 00:41:27.492 "medium_priority_weight": 0, 00:41:27.492 "high_priority_weight": 0, 00:41:27.492 "nvme_adminq_poll_period_us": 10000, 00:41:27.492 "nvme_ioq_poll_period_us": 0, 00:41:27.492 "io_queue_requests": 512, 00:41:27.492 "delay_cmd_submit": true, 00:41:27.492 "transport_retry_count": 4, 00:41:27.492 "bdev_retry_count": 3, 00:41:27.492 "transport_ack_timeout": 0, 00:41:27.492 "ctrlr_loss_timeout_sec": 0, 00:41:27.492 "reconnect_delay_sec": 0, 00:41:27.492 "fast_io_fail_timeout_sec": 0, 00:41:27.492 "disable_auto_failback": false, 00:41:27.492 "generate_uuids": false, 00:41:27.492 "transport_tos": 0, 00:41:27.492 "nvme_error_stat": false, 00:41:27.492 "rdma_srq_size": 0, 00:41:27.492 "io_path_stat": false, 00:41:27.492 "allow_accel_sequence": false, 00:41:27.492 "rdma_max_cq_size": 0, 00:41:27.492 "rdma_cm_event_timeout_ms": 0, 00:41:27.492 "dhchap_digests": [ 00:41:27.492 "sha256", 00:41:27.492 "sha384", 00:41:27.492 "sha512" 00:41:27.492 ], 00:41:27.492 "dhchap_dhgroups": [ 00:41:27.492 "null", 00:41:27.492 "ffdhe2048", 00:41:27.492 "ffdhe3072", 00:41:27.492 "ffdhe4096", 00:41:27.492 "ffdhe6144", 00:41:27.492 "ffdhe8192" 00:41:27.492 ] 00:41:27.492 } 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "method": "bdev_nvme_attach_controller", 00:41:27.492 "params": { 00:41:27.492 "name": "nvme0", 00:41:27.492 "trtype": "TCP", 00:41:27.492 "adrfam": "IPv4", 00:41:27.492 "traddr": "127.0.0.1", 00:41:27.492 "trsvcid": "4420", 00:41:27.492 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:41:27.492 "prchk_reftag": false, 00:41:27.492 "prchk_guard": false, 00:41:27.492 "ctrlr_loss_timeout_sec": 0, 00:41:27.492 "reconnect_delay_sec": 0, 00:41:27.492 "fast_io_fail_timeout_sec": 0, 00:41:27.492 "psk": "key0", 00:41:27.492 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:41:27.492 "hdgst": false, 00:41:27.492 "ddgst": false, 00:41:27.492 "multipath": "multipath" 00:41:27.492 } 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "method": "bdev_nvme_set_hotplug", 00:41:27.492 "params": { 00:41:27.492 "period_us": 100000, 00:41:27.492 "enable": false 00:41:27.492 } 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "method": "bdev_wait_for_examine" 00:41:27.492 } 00:41:27.492 ] 00:41:27.492 }, 00:41:27.492 { 00:41:27.492 "subsystem": "nbd", 00:41:27.492 "config": [] 00:41:27.492 } 00:41:27.492 ] 00:41:27.492 }' 00:41:27.492 09:17:17 keyring_file -- keyring/file.sh@115 -- # killprocess 2222292 00:41:27.492 09:17:17 keyring_file -- common/autotest_common.sh@950 -- # '[' -z 2222292 ']' 00:41:27.492 09:17:17 keyring_file -- common/autotest_common.sh@954 -- # kill -0 2222292 00:41:27.492 09:17:17 keyring_file -- common/autotest_common.sh@955 -- # uname 00:41:27.492 09:17:17 keyring_file -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:41:27.492 09:17:17 keyring_file -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2222292 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2222292' 00:41:27.753 killing process with pid 2222292 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@969 -- # kill 2222292 00:41:27.753 Received shutdown signal, test time was about 1.000000 seconds 00:41:27.753 00:41:27.753 Latency(us) 00:41:27.753 [2024-11-06T08:17:17.866Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:27.753 [2024-11-06T08:17:17.866Z] =================================================================================================================== 00:41:27.753 [2024-11-06T08:17:17.866Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@974 -- # wait 2222292 00:41:27.753 09:17:17 keyring_file -- keyring/file.sh@118 -- # bperfpid=2224095 00:41:27.753 09:17:17 keyring_file -- keyring/file.sh@120 -- # waitforlisten 2224095 /var/tmp/bperf.sock 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@831 -- # '[' -z 2224095 ']' 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@836 -- # local max_retries=100 00:41:27.753 09:17:17 keyring_file -- keyring/file.sh@116 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z -c /dev/fd/63 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:41:27.753 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@840 -- # xtrace_disable 00:41:27.753 09:17:17 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:41:27.753 09:17:17 keyring_file -- keyring/file.sh@116 -- # echo '{ 00:41:27.753 "subsystems": [ 00:41:27.753 { 00:41:27.753 "subsystem": "keyring", 00:41:27.753 "config": [ 00:41:27.753 { 00:41:27.753 "method": "keyring_file_add_key", 00:41:27.753 "params": { 00:41:27.753 "name": "key0", 00:41:27.753 "path": "/tmp/tmp.aEuhojhxs3" 00:41:27.753 } 00:41:27.753 }, 00:41:27.753 { 00:41:27.753 "method": "keyring_file_add_key", 00:41:27.753 "params": { 00:41:27.753 "name": "key1", 00:41:27.753 "path": "/tmp/tmp.wQtzL8FpXn" 00:41:27.753 } 00:41:27.753 } 00:41:27.753 ] 00:41:27.753 }, 00:41:27.753 { 00:41:27.753 "subsystem": "iobuf", 00:41:27.753 "config": [ 00:41:27.753 { 00:41:27.753 "method": "iobuf_set_options", 00:41:27.753 "params": { 00:41:27.753 "small_pool_count": 8192, 00:41:27.753 "large_pool_count": 1024, 00:41:27.753 "small_bufsize": 8192, 00:41:27.753 "large_bufsize": 135168, 00:41:27.753 "enable_numa": false 00:41:27.753 } 00:41:27.753 } 00:41:27.753 ] 00:41:27.753 }, 00:41:27.753 { 00:41:27.753 "subsystem": "sock", 00:41:27.753 "config": [ 00:41:27.753 { 00:41:27.753 "method": "sock_set_default_impl", 00:41:27.753 "params": { 00:41:27.753 "impl_name": "posix" 00:41:27.753 } 00:41:27.753 }, 00:41:27.754 { 00:41:27.754 "method": "sock_impl_set_options", 00:41:27.754 "params": { 00:41:27.754 "impl_name": "ssl", 00:41:27.754 "recv_buf_size": 4096, 00:41:27.754 "send_buf_size": 4096, 00:41:27.754 "enable_recv_pipe": true, 00:41:27.754 "enable_quickack": false, 00:41:27.754 "enable_placement_id": 0, 00:41:27.754 "enable_zerocopy_send_server": true, 00:41:27.754 "enable_zerocopy_send_client": false, 00:41:27.754 "zerocopy_threshold": 0, 00:41:27.754 "tls_version": 0, 00:41:27.754 "enable_ktls": false 00:41:27.754 } 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "method": "sock_impl_set_options", 00:41:27.754 "params": { 00:41:27.754 "impl_name": "posix", 00:41:27.754 "recv_buf_size": 2097152, 00:41:27.754 "send_buf_size": 2097152, 00:41:27.754 "enable_recv_pipe": true, 00:41:27.754 "enable_quickack": false, 00:41:27.754 "enable_placement_id": 0, 00:41:27.754 "enable_zerocopy_send_server": true, 00:41:27.754 "enable_zerocopy_send_client": false, 00:41:27.754 "zerocopy_threshold": 0, 00:41:27.754 "tls_version": 0, 00:41:27.754 "enable_ktls": false 00:41:27.754 } 00:41:27.754 } 00:41:27.754 ] 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "subsystem": "vmd", 00:41:27.754 "config": [] 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "subsystem": "accel", 00:41:27.754 "config": [ 00:41:27.754 { 00:41:27.754 "method": "accel_set_options", 00:41:27.754 "params": { 00:41:27.754 "small_cache_size": 128, 00:41:27.754 "large_cache_size": 16, 00:41:27.754 "task_count": 2048, 00:41:27.754 "sequence_count": 2048, 00:41:27.754 "buf_count": 2048 00:41:27.754 } 00:41:27.754 } 00:41:27.754 ] 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "subsystem": "bdev", 00:41:27.754 "config": [ 00:41:27.754 { 00:41:27.754 "method": "bdev_set_options", 00:41:27.754 "params": { 00:41:27.754 "bdev_io_pool_size": 65535, 00:41:27.754 "bdev_io_cache_size": 256, 00:41:27.754 "bdev_auto_examine": true, 00:41:27.754 "iobuf_small_cache_size": 128, 00:41:27.754 "iobuf_large_cache_size": 16 00:41:27.754 } 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "method": "bdev_raid_set_options", 00:41:27.754 "params": { 00:41:27.754 "process_window_size_kb": 1024, 00:41:27.754 "process_max_bandwidth_mb_sec": 0 00:41:27.754 } 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "method": "bdev_iscsi_set_options", 00:41:27.754 "params": { 00:41:27.754 "timeout_sec": 30 00:41:27.754 } 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "method": "bdev_nvme_set_options", 00:41:27.754 "params": { 00:41:27.754 "action_on_timeout": "none", 00:41:27.754 "timeout_us": 0, 00:41:27.754 "timeout_admin_us": 0, 00:41:27.754 "keep_alive_timeout_ms": 10000, 00:41:27.754 "arbitration_burst": 0, 00:41:27.754 "low_priority_weight": 0, 00:41:27.754 "medium_priority_weight": 0, 00:41:27.754 "high_priority_weight": 0, 00:41:27.754 "nvme_adminq_poll_period_us": 10000, 00:41:27.754 "nvme_ioq_poll_period_us": 0, 00:41:27.754 "io_queue_requests": 512, 00:41:27.754 "delay_cmd_submit": true, 00:41:27.754 "transport_retry_count": 4, 00:41:27.754 "bdev_retry_count": 3, 00:41:27.754 "transport_ack_timeout": 0, 00:41:27.754 "ctrlr_loss_timeout_sec": 0, 00:41:27.754 "reconnect_delay_sec": 0, 00:41:27.754 "fast_io_fail_timeout_sec": 0, 00:41:27.754 "disable_auto_failback": false, 00:41:27.754 "generate_uuids": false, 00:41:27.754 "transport_tos": 0, 00:41:27.754 "nvme_error_stat": false, 00:41:27.754 "rdma_srq_size": 0, 00:41:27.754 "io_path_stat": false, 00:41:27.754 "allow_accel_sequence": false, 00:41:27.754 "rdma_max_cq_size": 0, 00:41:27.754 "rdma_cm_event_timeout_ms": 0, 00:41:27.754 "dhchap_digests": [ 00:41:27.754 "sha256", 00:41:27.754 "sha384", 00:41:27.754 "sha512" 00:41:27.754 ], 00:41:27.754 "dhchap_dhgroups": [ 00:41:27.754 "null", 00:41:27.754 "ffdhe2048", 00:41:27.754 "ffdhe3072", 00:41:27.754 "ffdhe4096", 00:41:27.754 "ffdhe6144", 00:41:27.754 "ffdhe8192" 00:41:27.754 ] 00:41:27.754 } 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "method": "bdev_nvme_attach_controller", 00:41:27.754 "params": { 00:41:27.754 "name": "nvme0", 00:41:27.754 "trtype": "TCP", 00:41:27.754 "adrfam": "IPv4", 00:41:27.754 "traddr": "127.0.0.1", 00:41:27.754 "trsvcid": "4420", 00:41:27.754 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:41:27.754 "prchk_reftag": false, 00:41:27.754 "prchk_guard": false, 00:41:27.754 "ctrlr_loss_timeout_sec": 0, 00:41:27.754 "reconnect_delay_sec": 0, 00:41:27.754 "fast_io_fail_timeout_sec": 0, 00:41:27.754 "psk": "key0", 00:41:27.754 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:41:27.754 "hdgst": false, 00:41:27.754 "ddgst": false, 00:41:27.754 "multipath": "multipath" 00:41:27.754 } 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "method": "bdev_nvme_set_hotplug", 00:41:27.754 "params": { 00:41:27.754 "period_us": 100000, 00:41:27.754 "enable": false 00:41:27.754 } 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "method": "bdev_wait_for_examine" 00:41:27.754 } 00:41:27.754 ] 00:41:27.754 }, 00:41:27.754 { 00:41:27.754 "subsystem": "nbd", 00:41:27.754 "config": [] 00:41:27.754 } 00:41:27.754 ] 00:41:27.754 }' 00:41:27.754 [2024-11-06 09:17:17.817766] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:41:27.754 [2024-11-06 09:17:17.817822] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2224095 ] 00:41:28.014 [2024-11-06 09:17:17.900952] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:28.014 [2024-11-06 09:17:17.930415] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:41:28.014 [2024-11-06 09:17:18.074172] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:41:28.584 09:17:18 keyring_file -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:41:28.584 09:17:18 keyring_file -- common/autotest_common.sh@864 -- # return 0 00:41:28.584 09:17:18 keyring_file -- keyring/file.sh@121 -- # bperf_cmd keyring_get_keys 00:41:28.584 09:17:18 keyring_file -- keyring/file.sh@121 -- # jq length 00:41:28.584 09:17:18 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:28.844 09:17:18 keyring_file -- keyring/file.sh@121 -- # (( 2 == 2 )) 00:41:28.844 09:17:18 keyring_file -- keyring/file.sh@122 -- # get_refcnt key0 00:41:28.844 09:17:18 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:41:28.844 09:17:18 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:28.844 09:17:18 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:28.844 09:17:18 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:41:28.844 09:17:18 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:29.103 09:17:18 keyring_file -- keyring/file.sh@122 -- # (( 2 == 2 )) 00:41:29.103 09:17:18 keyring_file -- keyring/file.sh@123 -- # get_refcnt key1 00:41:29.103 09:17:18 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:41:29.103 09:17:18 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:41:29.103 09:17:18 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:29.103 09:17:18 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:41:29.103 09:17:18 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:29.103 09:17:19 keyring_file -- keyring/file.sh@123 -- # (( 1 == 1 )) 00:41:29.103 09:17:19 keyring_file -- keyring/file.sh@124 -- # bperf_cmd bdev_nvme_get_controllers 00:41:29.103 09:17:19 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_get_controllers 00:41:29.103 09:17:19 keyring_file -- keyring/file.sh@124 -- # jq -r '.[].name' 00:41:29.362 09:17:19 keyring_file -- keyring/file.sh@124 -- # [[ nvme0 == nvme0 ]] 00:41:29.362 09:17:19 keyring_file -- keyring/file.sh@1 -- # cleanup 00:41:29.362 09:17:19 keyring_file -- keyring/file.sh@19 -- # rm -f /tmp/tmp.aEuhojhxs3 /tmp/tmp.wQtzL8FpXn 00:41:29.362 09:17:19 keyring_file -- keyring/file.sh@20 -- # killprocess 2224095 00:41:29.362 09:17:19 keyring_file -- common/autotest_common.sh@950 -- # '[' -z 2224095 ']' 00:41:29.362 09:17:19 keyring_file -- common/autotest_common.sh@954 -- # kill -0 2224095 00:41:29.362 09:17:19 keyring_file -- common/autotest_common.sh@955 -- # uname 00:41:29.362 09:17:19 keyring_file -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:41:29.362 09:17:19 keyring_file -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2224095 00:41:29.362 09:17:19 keyring_file -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:41:29.362 09:17:19 keyring_file -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:41:29.362 09:17:19 keyring_file -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2224095' 00:41:29.362 killing process with pid 2224095 00:41:29.362 09:17:19 keyring_file -- common/autotest_common.sh@969 -- # kill 2224095 00:41:29.362 Received shutdown signal, test time was about 1.000000 seconds 00:41:29.362 00:41:29.362 Latency(us) 00:41:29.362 [2024-11-06T08:17:19.475Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:29.362 [2024-11-06T08:17:19.475Z] =================================================================================================================== 00:41:29.362 [2024-11-06T08:17:19.475Z] Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:41:29.362 09:17:19 keyring_file -- common/autotest_common.sh@974 -- # wait 2224095 00:41:29.636 09:17:19 keyring_file -- keyring/file.sh@21 -- # killprocess 2222274 00:41:29.636 09:17:19 keyring_file -- common/autotest_common.sh@950 -- # '[' -z 2222274 ']' 00:41:29.636 09:17:19 keyring_file -- common/autotest_common.sh@954 -- # kill -0 2222274 00:41:29.636 09:17:19 keyring_file -- common/autotest_common.sh@955 -- # uname 00:41:29.636 09:17:19 keyring_file -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:41:29.636 09:17:19 keyring_file -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2222274 00:41:29.636 09:17:19 keyring_file -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:41:29.636 09:17:19 keyring_file -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:41:29.636 09:17:19 keyring_file -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2222274' 00:41:29.636 killing process with pid 2222274 00:41:29.636 09:17:19 keyring_file -- common/autotest_common.sh@969 -- # kill 2222274 00:41:29.636 09:17:19 keyring_file -- common/autotest_common.sh@974 -- # wait 2222274 00:41:29.954 00:41:29.954 real 0m11.705s 00:41:29.954 user 0m28.158s 00:41:29.954 sys 0m2.532s 00:41:29.954 09:17:19 keyring_file -- common/autotest_common.sh@1126 -- # xtrace_disable 00:41:29.954 09:17:19 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:41:29.954 ************************************ 00:41:29.954 END TEST keyring_file 00:41:29.954 ************************************ 00:41:29.954 09:17:19 -- spdk/autotest.sh@289 -- # [[ y == y ]] 00:41:29.954 09:17:19 -- spdk/autotest.sh@290 -- # run_test keyring_linux /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/keyctl-session-wrapper /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/linux.sh 00:41:29.954 09:17:19 -- common/autotest_common.sh@1101 -- # '[' 3 -le 1 ']' 00:41:29.954 09:17:19 -- common/autotest_common.sh@1107 -- # xtrace_disable 00:41:29.954 09:17:19 -- common/autotest_common.sh@10 -- # set +x 00:41:29.954 ************************************ 00:41:29.954 START TEST keyring_linux 00:41:29.954 ************************************ 00:41:29.954 09:17:19 keyring_linux -- common/autotest_common.sh@1125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/keyctl-session-wrapper /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/linux.sh 00:41:29.954 Joined session keyring: 408727933 00:41:29.954 * Looking for test storage... 00:41:29.954 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:41:29.954 09:17:19 keyring_linux -- common/autotest_common.sh@1688 -- # [[ y == y ]] 00:41:29.954 09:17:19 keyring_linux -- common/autotest_common.sh@1689 -- # lcov --version 00:41:29.954 09:17:19 keyring_linux -- common/autotest_common.sh@1689 -- # awk '{print $NF}' 00:41:29.954 09:17:20 keyring_linux -- common/autotest_common.sh@1689 -- # lt 1.15 2 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@373 -- # cmp_versions 1.15 '<' 2 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@333 -- # local ver1 ver1_l 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@334 -- # local ver2 ver2_l 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@336 -- # IFS=.-: 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@336 -- # read -ra ver1 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@337 -- # IFS=.-: 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@337 -- # read -ra ver2 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@338 -- # local 'op=<' 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@340 -- # ver1_l=2 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@341 -- # ver2_l=1 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@343 -- # local lt=0 gt=0 eq=0 v 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@344 -- # case "$op" in 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@345 -- # : 1 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@364 -- # (( v = 0 )) 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@364 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@365 -- # decimal 1 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@353 -- # local d=1 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@354 -- # [[ 1 =~ ^[0-9]+$ ]] 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@355 -- # echo 1 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@365 -- # ver1[v]=1 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@366 -- # decimal 2 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@353 -- # local d=2 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@354 -- # [[ 2 =~ ^[0-9]+$ ]] 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@355 -- # echo 2 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@366 -- # ver2[v]=2 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@367 -- # (( ver1[v] > ver2[v] )) 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@368 -- # (( ver1[v] < ver2[v] )) 00:41:29.954 09:17:20 keyring_linux -- scripts/common.sh@368 -- # return 0 00:41:29.954 09:17:20 keyring_linux -- common/autotest_common.sh@1690 -- # lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:41:29.954 09:17:20 keyring_linux -- common/autotest_common.sh@1702 -- # export 'LCOV_OPTS= 00:41:29.954 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:29.954 --rc genhtml_branch_coverage=1 00:41:29.954 --rc genhtml_function_coverage=1 00:41:29.954 --rc genhtml_legend=1 00:41:29.954 --rc geninfo_all_blocks=1 00:41:29.954 --rc geninfo_unexecuted_blocks=1 00:41:29.954 00:41:29.954 ' 00:41:29.954 09:17:20 keyring_linux -- common/autotest_common.sh@1702 -- # LCOV_OPTS=' 00:41:29.954 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:29.954 --rc genhtml_branch_coverage=1 00:41:29.954 --rc genhtml_function_coverage=1 00:41:29.954 --rc genhtml_legend=1 00:41:29.954 --rc geninfo_all_blocks=1 00:41:29.954 --rc geninfo_unexecuted_blocks=1 00:41:29.954 00:41:29.954 ' 00:41:29.954 09:17:20 keyring_linux -- common/autotest_common.sh@1703 -- # export 'LCOV=lcov 00:41:29.954 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:29.954 --rc genhtml_branch_coverage=1 00:41:29.954 --rc genhtml_function_coverage=1 00:41:29.954 --rc genhtml_legend=1 00:41:29.954 --rc geninfo_all_blocks=1 00:41:29.954 --rc geninfo_unexecuted_blocks=1 00:41:29.954 00:41:29.954 ' 00:41:29.954 09:17:20 keyring_linux -- common/autotest_common.sh@1703 -- # LCOV='lcov 00:41:29.954 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:41:29.954 --rc genhtml_branch_coverage=1 00:41:29.954 --rc genhtml_function_coverage=1 00:41:29.954 --rc genhtml_legend=1 00:41:29.954 --rc geninfo_all_blocks=1 00:41:29.954 --rc geninfo_unexecuted_blocks=1 00:41:29.954 00:41:29.954 ' 00:41:29.954 09:17:20 keyring_linux -- keyring/linux.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:41:29.954 09:17:20 keyring_linux -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@7 -- # uname -s 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:41:29.954 09:17:20 keyring_linux -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:00d0226a-fbea-ec11-9bc7-a4bf019282be 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@18 -- # NVME_HOSTID=00d0226a-fbea-ec11-9bc7-a4bf019282be 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@49 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:41:30.244 09:17:20 keyring_linux -- scripts/common.sh@15 -- # shopt -s extglob 00:41:30.244 09:17:20 keyring_linux -- scripts/common.sh@544 -- # [[ -e /bin/wpdk_common.sh ]] 00:41:30.244 09:17:20 keyring_linux -- scripts/common.sh@552 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:41:30.244 09:17:20 keyring_linux -- scripts/common.sh@553 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:41:30.244 09:17:20 keyring_linux -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:30.244 09:17:20 keyring_linux -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:30.244 09:17:20 keyring_linux -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:30.244 09:17:20 keyring_linux -- paths/export.sh@5 -- # export PATH 00:41:30.244 09:17:20 keyring_linux -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@51 -- # : 0 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@52 -- # export NVMF_APP_SHM_ID 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@53 -- # build_nvmf_app_args 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@33 -- # '[' '' -eq 1 ']' 00:41:30.244 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh: line 33: [: : integer expression expected 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@37 -- # '[' -n '' ']' 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@39 -- # '[' 0 -eq 1 ']' 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@55 -- # have_pci_nics=0 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:41:30.244 09:17:20 keyring_linux -- keyring/linux.sh@11 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:41:30.244 09:17:20 keyring_linux -- keyring/linux.sh@12 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:41:30.244 09:17:20 keyring_linux -- keyring/linux.sh@13 -- # key0=00112233445566778899aabbccddeeff 00:41:30.244 09:17:20 keyring_linux -- keyring/linux.sh@14 -- # key1=112233445566778899aabbccddeeff00 00:41:30.244 09:17:20 keyring_linux -- keyring/linux.sh@45 -- # trap cleanup EXIT 00:41:30.244 09:17:20 keyring_linux -- keyring/linux.sh@47 -- # prep_key key0 00112233445566778899aabbccddeeff 0 /tmp/:spdk-test:key0 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@15 -- # local name key digest path 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@17 -- # name=key0 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@17 -- # digest=0 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@18 -- # path=/tmp/:spdk-test:key0 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@728 -- # local prefix key digest 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@730 -- # key=00112233445566778899aabbccddeeff 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@730 -- # digest=0 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@731 -- # python - 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@21 -- # chmod 0600 /tmp/:spdk-test:key0 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@23 -- # echo /tmp/:spdk-test:key0 00:41:30.244 /tmp/:spdk-test:key0 00:41:30.244 09:17:20 keyring_linux -- keyring/linux.sh@48 -- # prep_key key1 112233445566778899aabbccddeeff00 0 /tmp/:spdk-test:key1 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@15 -- # local name key digest path 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@17 -- # name=key1 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@17 -- # digest=0 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@18 -- # path=/tmp/:spdk-test:key1 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@741 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@728 -- # local prefix key digest 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@730 -- # prefix=NVMeTLSkey-1 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@730 -- # key=112233445566778899aabbccddeeff00 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@730 -- # digest=0 00:41:30.244 09:17:20 keyring_linux -- nvmf/common.sh@731 -- # python - 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@21 -- # chmod 0600 /tmp/:spdk-test:key1 00:41:30.244 09:17:20 keyring_linux -- keyring/common.sh@23 -- # echo /tmp/:spdk-test:key1 00:41:30.244 /tmp/:spdk-test:key1 00:41:30.244 09:17:20 keyring_linux -- keyring/linux.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:41:30.244 09:17:20 keyring_linux -- keyring/linux.sh@51 -- # tgtpid=2224536 00:41:30.244 09:17:20 keyring_linux -- keyring/linux.sh@53 -- # waitforlisten 2224536 00:41:30.244 09:17:20 keyring_linux -- common/autotest_common.sh@831 -- # '[' -z 2224536 ']' 00:41:30.244 09:17:20 keyring_linux -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/spdk.sock 00:41:30.244 09:17:20 keyring_linux -- common/autotest_common.sh@836 -- # local max_retries=100 00:41:30.244 09:17:20 keyring_linux -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:41:30.244 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:41:30.244 09:17:20 keyring_linux -- common/autotest_common.sh@840 -- # xtrace_disable 00:41:30.244 09:17:20 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:41:30.244 [2024-11-06 09:17:20.239072] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:41:30.244 [2024-11-06 09:17:20.239145] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2224536 ] 00:41:30.244 [2024-11-06 09:17:20.310554] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:30.550 [2024-11-06 09:17:20.346554] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 0 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@864 -- # return 0 00:41:31.126 09:17:21 keyring_linux -- keyring/linux.sh@54 -- # rpc_cmd 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@561 -- # xtrace_disable 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:41:31.126 [2024-11-06 09:17:21.010777] tcp.c: 738:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:41:31.126 null0 00:41:31.126 [2024-11-06 09:17:21.042821] tcp.c:1031:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:41:31.126 [2024-11-06 09:17:21.043221] tcp.c:1081:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@589 -- # [[ 0 == 0 ]] 00:41:31.126 09:17:21 keyring_linux -- keyring/linux.sh@66 -- # keyctl add user :spdk-test:key0 NVMeTLSkey-1:00:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: @s 00:41:31.126 809716731 00:41:31.126 09:17:21 keyring_linux -- keyring/linux.sh@67 -- # keyctl add user :spdk-test:key1 NVMeTLSkey-1:00:MTEyMjMzNDQ1NTY2Nzc4ODk5YWFiYmNjZGRlZWZmMDA6CPcs: @s 00:41:31.126 731801113 00:41:31.126 09:17:21 keyring_linux -- keyring/linux.sh@70 -- # bperfpid=2224872 00:41:31.126 09:17:21 keyring_linux -- keyring/linux.sh@72 -- # waitforlisten 2224872 /var/tmp/bperf.sock 00:41:31.126 09:17:21 keyring_linux -- keyring/linux.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randread -t 1 -m 2 -r /var/tmp/bperf.sock -z --wait-for-rpc 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@831 -- # '[' -z 2224872 ']' 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@835 -- # local rpc_addr=/var/tmp/bperf.sock 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@836 -- # local max_retries=100 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@838 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:41:31.126 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@840 -- # xtrace_disable 00:41:31.126 09:17:21 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:41:31.126 [2024-11-06 09:17:21.131095] Starting SPDK v25.01-pre git sha1 15b283ee8 / DPDK 24.03.0 initialization... 00:41:31.126 [2024-11-06 09:17:21.131143] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid2224872 ] 00:41:31.126 [2024-11-06 09:17:21.213737] app.c: 919:spdk_app_start: *NOTICE*: Total cores available: 1 00:41:31.388 [2024-11-06 09:17:21.243399] reactor.c:1005:reactor_run: *NOTICE*: Reactor started on core 1 00:41:31.959 09:17:21 keyring_linux -- common/autotest_common.sh@860 -- # (( i == 0 )) 00:41:31.959 09:17:21 keyring_linux -- common/autotest_common.sh@864 -- # return 0 00:41:31.959 09:17:21 keyring_linux -- keyring/linux.sh@73 -- # bperf_cmd keyring_linux_set_options --enable 00:41:31.959 09:17:21 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_linux_set_options --enable 00:41:31.959 09:17:22 keyring_linux -- keyring/linux.sh@74 -- # bperf_cmd framework_start_init 00:41:31.959 09:17:22 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:41:32.221 09:17:22 keyring_linux -- keyring/linux.sh@75 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key0 00:41:32.221 09:17:22 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key0 00:41:32.481 [2024-11-06 09:17:22.448657] bdev_nvme_rpc.c: 514:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:41:32.481 nvme0n1 00:41:32.481 09:17:22 keyring_linux -- keyring/linux.sh@77 -- # check_keys 1 :spdk-test:key0 00:41:32.481 09:17:22 keyring_linux -- keyring/linux.sh@19 -- # local count=1 name=:spdk-test:key0 00:41:32.481 09:17:22 keyring_linux -- keyring/linux.sh@20 -- # local sn 00:41:32.481 09:17:22 keyring_linux -- keyring/linux.sh@22 -- # bperf_cmd keyring_get_keys 00:41:32.481 09:17:22 keyring_linux -- keyring/linux.sh@22 -- # jq length 00:41:32.481 09:17:22 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:32.743 09:17:22 keyring_linux -- keyring/linux.sh@22 -- # (( 1 == count )) 00:41:32.743 09:17:22 keyring_linux -- keyring/linux.sh@23 -- # (( count == 0 )) 00:41:32.743 09:17:22 keyring_linux -- keyring/linux.sh@25 -- # get_key :spdk-test:key0 00:41:32.743 09:17:22 keyring_linux -- keyring/linux.sh@25 -- # jq -r .sn 00:41:32.743 09:17:22 keyring_linux -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:41:32.743 09:17:22 keyring_linux -- keyring/common.sh@10 -- # jq '.[] | select(.name == ":spdk-test:key0")' 00:41:32.743 09:17:22 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:33.003 09:17:22 keyring_linux -- keyring/linux.sh@25 -- # sn=809716731 00:41:33.003 09:17:22 keyring_linux -- keyring/linux.sh@26 -- # get_keysn :spdk-test:key0 00:41:33.003 09:17:22 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key0 00:41:33.003 09:17:22 keyring_linux -- keyring/linux.sh@26 -- # [[ 809716731 == \8\0\9\7\1\6\7\3\1 ]] 00:41:33.003 09:17:22 keyring_linux -- keyring/linux.sh@27 -- # keyctl print 809716731 00:41:33.003 09:17:22 keyring_linux -- keyring/linux.sh@27 -- # [[ NVMeTLSkey-1:00:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: == \N\V\M\e\T\L\S\k\e\y\-\1\:\0\0\:\M\D\A\x\M\T\I\y\M\z\M\0\N\D\U\1\N\j\Y\3\N\z\g\4\O\T\l\h\Y\W\J\i\Y\2\N\k\Z\G\V\l\Z\m\Z\w\J\E\i\Q\: ]] 00:41:33.003 09:17:22 keyring_linux -- keyring/linux.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:41:33.003 Running I/O for 1 seconds... 00:41:33.945 16294.00 IOPS, 63.65 MiB/s 00:41:33.945 Latency(us) 00:41:33.945 [2024-11-06T08:17:24.058Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:33.945 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:41:33.945 nvme0n1 : 1.01 16294.71 63.65 0.00 0.00 7821.60 6225.92 13216.43 00:41:33.945 [2024-11-06T08:17:24.058Z] =================================================================================================================== 00:41:33.945 [2024-11-06T08:17:24.058Z] Total : 16294.71 63.65 0.00 0.00 7821.60 6225.92 13216.43 00:41:33.945 { 00:41:33.945 "results": [ 00:41:33.945 { 00:41:33.945 "job": "nvme0n1", 00:41:33.945 "core_mask": "0x2", 00:41:33.945 "workload": "randread", 00:41:33.945 "status": "finished", 00:41:33.945 "queue_depth": 128, 00:41:33.945 "io_size": 4096, 00:41:33.945 "runtime": 1.007812, 00:41:33.945 "iops": 16294.705758613709, 00:41:33.945 "mibps": 63.6511943695848, 00:41:33.945 "io_failed": 0, 00:41:33.945 "io_timeout": 0, 00:41:33.945 "avg_latency_us": 7821.595758535298, 00:41:33.945 "min_latency_us": 6225.92, 00:41:33.945 "max_latency_us": 13216.426666666666 00:41:33.945 } 00:41:33.945 ], 00:41:33.945 "core_count": 1 00:41:33.945 } 00:41:33.945 09:17:24 keyring_linux -- keyring/linux.sh@80 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:41:33.945 09:17:24 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:41:34.207 09:17:24 keyring_linux -- keyring/linux.sh@81 -- # check_keys 0 00:41:34.207 09:17:24 keyring_linux -- keyring/linux.sh@19 -- # local count=0 name= 00:41:34.207 09:17:24 keyring_linux -- keyring/linux.sh@20 -- # local sn 00:41:34.207 09:17:24 keyring_linux -- keyring/linux.sh@22 -- # bperf_cmd keyring_get_keys 00:41:34.207 09:17:24 keyring_linux -- keyring/linux.sh@22 -- # jq length 00:41:34.207 09:17:24 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:41:34.469 09:17:24 keyring_linux -- keyring/linux.sh@22 -- # (( 0 == count )) 00:41:34.469 09:17:24 keyring_linux -- keyring/linux.sh@23 -- # (( count == 0 )) 00:41:34.469 09:17:24 keyring_linux -- keyring/linux.sh@23 -- # return 00:41:34.469 09:17:24 keyring_linux -- keyring/linux.sh@84 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:41:34.469 09:17:24 keyring_linux -- common/autotest_common.sh@650 -- # local es=0 00:41:34.469 09:17:24 keyring_linux -- common/autotest_common.sh@652 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:41:34.469 09:17:24 keyring_linux -- common/autotest_common.sh@638 -- # local arg=bperf_cmd 00:41:34.469 09:17:24 keyring_linux -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:41:34.469 09:17:24 keyring_linux -- common/autotest_common.sh@642 -- # type -t bperf_cmd 00:41:34.469 09:17:24 keyring_linux -- common/autotest_common.sh@642 -- # case "$(type -t "$arg")" in 00:41:34.469 09:17:24 keyring_linux -- common/autotest_common.sh@653 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:41:34.469 09:17:24 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:41:34.469 [2024-11-06 09:17:24.513295] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 421:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:41:34.469 [2024-11-06 09:17:24.513917] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x21b89b0 (107): Transport endpoint is not connected 00:41:34.469 [2024-11-06 09:17:24.514914] nvme_tcp.c:2085:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x21b89b0 (9): Bad file descriptor 00:41:34.469 [2024-11-06 09:17:24.515916] nvme_ctrlr.c:4170:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] Ctrlr is in error state 00:41:34.470 [2024-11-06 09:17:24.515923] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:41:34.470 [2024-11-06 09:17:24.515928] nvme.c: 884:nvme_dummy_attach_fail_cb: *ERROR*: Failed to attach nvme ctrlr: trtype=TCP adrfam=IPv4 traddr=127.0.0.1 trsvcid=4420 subnqn=nqn.2016-06.io.spdk:cnode0, Operation not permitted 00:41:34.470 [2024-11-06 09:17:24.515935] nvme_ctrlr.c:1083:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0, 0] in failed state. 00:41:34.470 request: 00:41:34.470 { 00:41:34.470 "name": "nvme0", 00:41:34.470 "trtype": "tcp", 00:41:34.470 "traddr": "127.0.0.1", 00:41:34.470 "adrfam": "ipv4", 00:41:34.470 "trsvcid": "4420", 00:41:34.470 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:41:34.470 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:41:34.470 "prchk_reftag": false, 00:41:34.470 "prchk_guard": false, 00:41:34.470 "hdgst": false, 00:41:34.470 "ddgst": false, 00:41:34.470 "psk": ":spdk-test:key1", 00:41:34.470 "allow_unrecognized_csi": false, 00:41:34.470 "method": "bdev_nvme_attach_controller", 00:41:34.470 "req_id": 1 00:41:34.470 } 00:41:34.470 Got JSON-RPC error response 00:41:34.470 response: 00:41:34.470 { 00:41:34.470 "code": -5, 00:41:34.470 "message": "Input/output error" 00:41:34.470 } 00:41:34.470 09:17:24 keyring_linux -- common/autotest_common.sh@653 -- # es=1 00:41:34.470 09:17:24 keyring_linux -- common/autotest_common.sh@661 -- # (( es > 128 )) 00:41:34.470 09:17:24 keyring_linux -- common/autotest_common.sh@672 -- # [[ -n '' ]] 00:41:34.470 09:17:24 keyring_linux -- common/autotest_common.sh@677 -- # (( !es == 0 )) 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@1 -- # cleanup 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@38 -- # for key in key0 key1 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@39 -- # unlink_key key0 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@31 -- # local name=key0 sn 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@33 -- # get_keysn :spdk-test:key0 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key0 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@33 -- # sn=809716731 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@34 -- # keyctl unlink 809716731 00:41:34.470 1 links removed 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@38 -- # for key in key0 key1 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@39 -- # unlink_key key1 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@31 -- # local name=key1 sn 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@33 -- # get_keysn :spdk-test:key1 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key1 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@33 -- # sn=731801113 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@34 -- # keyctl unlink 731801113 00:41:34.470 1 links removed 00:41:34.470 09:17:24 keyring_linux -- keyring/linux.sh@41 -- # killprocess 2224872 00:41:34.470 09:17:24 keyring_linux -- common/autotest_common.sh@950 -- # '[' -z 2224872 ']' 00:41:34.470 09:17:24 keyring_linux -- common/autotest_common.sh@954 -- # kill -0 2224872 00:41:34.470 09:17:24 keyring_linux -- common/autotest_common.sh@955 -- # uname 00:41:34.470 09:17:24 keyring_linux -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:41:34.470 09:17:24 keyring_linux -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2224872 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@956 -- # process_name=reactor_1 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@960 -- # '[' reactor_1 = sudo ']' 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2224872' 00:41:34.731 killing process with pid 2224872 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@969 -- # kill 2224872 00:41:34.731 Received shutdown signal, test time was about 1.000000 seconds 00:41:34.731 00:41:34.731 Latency(us) 00:41:34.731 [2024-11-06T08:17:24.844Z] Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:41:34.731 [2024-11-06T08:17:24.844Z] =================================================================================================================== 00:41:34.731 [2024-11-06T08:17:24.844Z] Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@974 -- # wait 2224872 00:41:34.731 09:17:24 keyring_linux -- keyring/linux.sh@42 -- # killprocess 2224536 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@950 -- # '[' -z 2224536 ']' 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@954 -- # kill -0 2224536 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@955 -- # uname 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@955 -- # '[' Linux = Linux ']' 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@956 -- # ps --no-headers -o comm= 2224536 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@956 -- # process_name=reactor_0 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@960 -- # '[' reactor_0 = sudo ']' 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@968 -- # echo 'killing process with pid 2224536' 00:41:34.731 killing process with pid 2224536 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@969 -- # kill 2224536 00:41:34.731 09:17:24 keyring_linux -- common/autotest_common.sh@974 -- # wait 2224536 00:41:34.992 00:41:34.992 real 0m5.161s 00:41:34.992 user 0m9.534s 00:41:34.992 sys 0m1.396s 00:41:34.992 09:17:24 keyring_linux -- common/autotest_common.sh@1126 -- # xtrace_disable 00:41:34.992 09:17:24 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:41:34.992 ************************************ 00:41:34.992 END TEST keyring_linux 00:41:34.992 ************************************ 00:41:34.992 09:17:25 -- spdk/autotest.sh@307 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@311 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@315 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@320 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@329 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@334 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@338 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@342 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@346 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@351 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@355 -- # '[' 0 -eq 1 ']' 00:41:34.992 09:17:25 -- spdk/autotest.sh@362 -- # [[ 0 -eq 1 ]] 00:41:34.992 09:17:25 -- spdk/autotest.sh@366 -- # [[ 0 -eq 1 ]] 00:41:34.992 09:17:25 -- spdk/autotest.sh@370 -- # [[ 0 -eq 1 ]] 00:41:34.992 09:17:25 -- spdk/autotest.sh@374 -- # [[ '' -eq 1 ]] 00:41:34.992 09:17:25 -- spdk/autotest.sh@381 -- # trap - SIGINT SIGTERM EXIT 00:41:34.992 09:17:25 -- spdk/autotest.sh@383 -- # timing_enter post_cleanup 00:41:34.992 09:17:25 -- common/autotest_common.sh@724 -- # xtrace_disable 00:41:34.992 09:17:25 -- common/autotest_common.sh@10 -- # set +x 00:41:34.992 09:17:25 -- spdk/autotest.sh@384 -- # autotest_cleanup 00:41:34.992 09:17:25 -- common/autotest_common.sh@1392 -- # local autotest_es=0 00:41:34.992 09:17:25 -- common/autotest_common.sh@1393 -- # xtrace_disable 00:41:34.992 09:17:25 -- common/autotest_common.sh@10 -- # set +x 00:41:43.133 INFO: APP EXITING 00:41:43.133 INFO: killing all VMs 00:41:43.133 INFO: killing vhost app 00:41:43.134 INFO: EXIT DONE 00:41:45.678 0000:80:01.6 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:80:01.7 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:80:01.4 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:80:01.5 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:80:01.2 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:80:01.3 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:80:01.0 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:80:01.1 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:65:00.0 (144d a80a): Already using the nvme driver 00:41:45.678 0000:00:01.6 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:00:01.7 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:00:01.4 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:00:01.5 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:00:01.2 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:00:01.3 (8086 0b00): Already using the ioatdma driver 00:41:45.678 0000:00:01.0 (8086 0b00): Already using the ioatdma driver 00:41:45.679 0000:00:01.1 (8086 0b00): Already using the ioatdma driver 00:41:49.888 Cleaning 00:41:49.888 Removing: /var/run/dpdk/spdk0/config 00:41:49.888 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:41:49.888 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:41:49.888 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:41:49.888 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:41:49.888 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-0 00:41:49.888 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-1 00:41:49.888 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-2 00:41:49.888 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-3 00:41:49.888 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:41:49.888 Removing: /var/run/dpdk/spdk0/hugepage_info 00:41:49.888 Removing: /var/run/dpdk/spdk1/config 00:41:49.888 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-0 00:41:49.888 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-1 00:41:49.888 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-2 00:41:49.888 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-3 00:41:49.888 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-0 00:41:49.888 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-1 00:41:49.888 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-2 00:41:49.888 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-3 00:41:49.889 Removing: /var/run/dpdk/spdk1/fbarray_memzone 00:41:49.889 Removing: /var/run/dpdk/spdk1/hugepage_info 00:41:49.889 Removing: /var/run/dpdk/spdk2/config 00:41:49.889 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-0 00:41:49.889 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-1 00:41:49.889 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-2 00:41:49.889 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-3 00:41:49.889 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-0 00:41:49.889 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-1 00:41:49.889 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-2 00:41:49.889 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-3 00:41:49.889 Removing: /var/run/dpdk/spdk2/fbarray_memzone 00:41:49.889 Removing: /var/run/dpdk/spdk2/hugepage_info 00:41:49.889 Removing: /var/run/dpdk/spdk3/config 00:41:49.889 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-0 00:41:49.889 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-1 00:41:49.889 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-2 00:41:49.889 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-3 00:41:49.889 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-0 00:41:49.889 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-1 00:41:49.889 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-2 00:41:49.889 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-3 00:41:49.889 Removing: /var/run/dpdk/spdk3/fbarray_memzone 00:41:49.889 Removing: /var/run/dpdk/spdk3/hugepage_info 00:41:49.889 Removing: /var/run/dpdk/spdk4/config 00:41:49.889 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-0 00:41:49.889 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-1 00:41:49.889 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-2 00:41:49.889 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-3 00:41:49.889 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-0 00:41:49.889 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-1 00:41:49.889 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-2 00:41:49.889 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-3 00:41:49.889 Removing: /var/run/dpdk/spdk4/fbarray_memzone 00:41:49.889 Removing: /var/run/dpdk/spdk4/hugepage_info 00:41:49.889 Removing: /dev/shm/bdev_svc_trace.1 00:41:49.889 Removing: /dev/shm/nvmf_trace.0 00:41:49.889 Removing: /dev/shm/spdk_tgt_trace.pid1653176 00:41:49.889 Removing: /var/run/dpdk/spdk0 00:41:49.889 Removing: /var/run/dpdk/spdk1 00:41:49.889 Removing: /var/run/dpdk/spdk2 00:41:49.889 Removing: /var/run/dpdk/spdk3 00:41:49.889 Removing: /var/run/dpdk/spdk4 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1651570 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1653176 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1653803 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1655000 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1655148 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1656464 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1656498 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1656952 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1658091 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1658573 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1658952 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1659350 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1659763 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1660168 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1660519 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1660727 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1660997 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1662042 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1665591 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1665825 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1666000 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1666134 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1666700 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1666729 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1667322 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1667411 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1667647 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1667790 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1668134 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1668171 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1668748 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1668968 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1669375 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1674085 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1679841 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1691927 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1692609 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1697822 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1698308 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1703433 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1710312 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1713610 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1726112 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1737473 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1739687 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1740802 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1761459 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1766211 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1823419 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1829802 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1836982 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1845448 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1845453 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1846454 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1847459 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1848474 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1849370 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1849474 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1849768 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1849822 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1849824 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1850829 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1851838 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1852888 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1853534 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1853684 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1853950 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1855310 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1856691 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1866679 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1902843 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1908256 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1910254 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1912288 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1912603 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1912625 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1912906 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1913347 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1915388 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1916465 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1917145 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1919717 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1920244 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1921376 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1926723 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1933280 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1933281 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1933283 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1937979 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1948236 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1953122 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1960287 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1961830 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1963628 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1965144 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1970906 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1975900 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1985533 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1985600 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1990650 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1990990 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1991272 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1991668 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1991679 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1997363 00:41:49.889 Removing: /var/run/dpdk/spdk_pid1997909 00:41:49.889 Removing: /var/run/dpdk/spdk_pid2003380 00:41:49.889 Removing: /var/run/dpdk/spdk_pid2006524 00:41:49.889 Removing: /var/run/dpdk/spdk_pid2012996 00:41:49.889 Removing: /var/run/dpdk/spdk_pid2019399 00:41:49.889 Removing: /var/run/dpdk/spdk_pid2029563 00:41:49.889 Removing: /var/run/dpdk/spdk_pid2038498 00:41:49.889 Removing: /var/run/dpdk/spdk_pid2038544 00:41:49.889 Removing: /var/run/dpdk/spdk_pid2062013 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2062808 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2063662 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2064409 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2065151 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2065996 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2066801 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2067510 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2072616 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2072948 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2080021 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2080373 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2087420 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2092500 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2104082 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2104802 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2109850 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2110201 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2115231 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2121956 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2124822 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2137582 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2148124 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2150130 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2151142 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2170737 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2175453 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2178631 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2185849 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2185915 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2192344 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2194714 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2196944 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2198413 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2200739 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2202136 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2212087 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2212769 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2213435 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2216267 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2216738 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2217402 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2222274 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2222292 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2224095 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2224536 00:41:50.150 Removing: /var/run/dpdk/spdk_pid2224872 00:41:50.150 Clean 00:41:50.412 09:17:40 -- common/autotest_common.sh@1449 -- # return 0 00:41:50.412 09:17:40 -- spdk/autotest.sh@385 -- # timing_exit post_cleanup 00:41:50.412 09:17:40 -- common/autotest_common.sh@730 -- # xtrace_disable 00:41:50.412 09:17:40 -- common/autotest_common.sh@10 -- # set +x 00:41:50.412 09:17:40 -- spdk/autotest.sh@387 -- # timing_exit autotest 00:41:50.412 09:17:40 -- common/autotest_common.sh@730 -- # xtrace_disable 00:41:50.412 09:17:40 -- common/autotest_common.sh@10 -- # set +x 00:41:50.412 09:17:40 -- spdk/autotest.sh@388 -- # chmod a+r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:41:50.412 09:17:40 -- spdk/autotest.sh@390 -- # [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log ]] 00:41:50.412 09:17:40 -- spdk/autotest.sh@390 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log 00:41:50.412 09:17:40 -- spdk/autotest.sh@392 -- # [[ y == y ]] 00:41:50.412 09:17:40 -- spdk/autotest.sh@394 -- # hostname 00:41:50.412 09:17:40 -- spdk/autotest.sh@394 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -c --no-external -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -t spdk-cyp-09 -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info 00:41:50.672 geninfo: WARNING: invalid characters removed from testname! 00:42:17.252 09:18:05 -- spdk/autotest.sh@395 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:42:18.634 09:18:08 -- spdk/autotest.sh@396 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/dpdk/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:42:20.546 09:18:10 -- spdk/autotest.sh@400 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info --ignore-errors unused,unused '/usr/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:42:22.459 09:18:12 -- spdk/autotest.sh@401 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/examples/vmd/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:42:23.842 09:18:13 -- spdk/autotest.sh@402 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:42:25.755 09:18:15 -- spdk/autotest.sh@403 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --rc geninfo_unexecuted_blocks=1 -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:42:27.139 09:18:16 -- spdk/autotest.sh@404 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:42:27.139 09:18:17 -- common/autotest_common.sh@1688 -- $ [[ y == y ]] 00:42:27.139 09:18:17 -- common/autotest_common.sh@1689 -- $ awk '{print $NF}' 00:42:27.139 09:18:17 -- common/autotest_common.sh@1689 -- $ lcov --version 00:42:27.139 09:18:17 -- common/autotest_common.sh@1689 -- $ lt 1.15 2 00:42:27.139 09:18:17 -- scripts/common.sh@373 -- $ cmp_versions 1.15 '<' 2 00:42:27.139 09:18:17 -- scripts/common.sh@333 -- $ local ver1 ver1_l 00:42:27.139 09:18:17 -- scripts/common.sh@334 -- $ local ver2 ver2_l 00:42:27.139 09:18:17 -- scripts/common.sh@336 -- $ IFS=.-: 00:42:27.139 09:18:17 -- scripts/common.sh@336 -- $ read -ra ver1 00:42:27.139 09:18:17 -- scripts/common.sh@337 -- $ IFS=.-: 00:42:27.139 09:18:17 -- scripts/common.sh@337 -- $ read -ra ver2 00:42:27.139 09:18:17 -- scripts/common.sh@338 -- $ local 'op=<' 00:42:27.139 09:18:17 -- scripts/common.sh@340 -- $ ver1_l=2 00:42:27.139 09:18:17 -- scripts/common.sh@341 -- $ ver2_l=1 00:42:27.139 09:18:17 -- scripts/common.sh@343 -- $ local lt=0 gt=0 eq=0 v 00:42:27.139 09:18:17 -- scripts/common.sh@344 -- $ case "$op" in 00:42:27.139 09:18:17 -- scripts/common.sh@345 -- $ : 1 00:42:27.139 09:18:17 -- scripts/common.sh@364 -- $ (( v = 0 )) 00:42:27.139 09:18:17 -- scripts/common.sh@364 -- $ (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:42:27.139 09:18:17 -- scripts/common.sh@365 -- $ decimal 1 00:42:27.139 09:18:17 -- scripts/common.sh@353 -- $ local d=1 00:42:27.139 09:18:17 -- scripts/common.sh@354 -- $ [[ 1 =~ ^[0-9]+$ ]] 00:42:27.139 09:18:17 -- scripts/common.sh@355 -- $ echo 1 00:42:27.139 09:18:17 -- scripts/common.sh@365 -- $ ver1[v]=1 00:42:27.139 09:18:17 -- scripts/common.sh@366 -- $ decimal 2 00:42:27.139 09:18:17 -- scripts/common.sh@353 -- $ local d=2 00:42:27.139 09:18:17 -- scripts/common.sh@354 -- $ [[ 2 =~ ^[0-9]+$ ]] 00:42:27.139 09:18:17 -- scripts/common.sh@355 -- $ echo 2 00:42:27.139 09:18:17 -- scripts/common.sh@366 -- $ ver2[v]=2 00:42:27.139 09:18:17 -- scripts/common.sh@367 -- $ (( ver1[v] > ver2[v] )) 00:42:27.139 09:18:17 -- scripts/common.sh@368 -- $ (( ver1[v] < ver2[v] )) 00:42:27.139 09:18:17 -- scripts/common.sh@368 -- $ return 0 00:42:27.139 09:18:17 -- common/autotest_common.sh@1690 -- $ lcov_rc_opt='--rc lcov_branch_coverage=1 --rc lcov_function_coverage=1' 00:42:27.139 09:18:17 -- common/autotest_common.sh@1702 -- $ export 'LCOV_OPTS= 00:42:27.139 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:27.139 --rc genhtml_branch_coverage=1 00:42:27.139 --rc genhtml_function_coverage=1 00:42:27.139 --rc genhtml_legend=1 00:42:27.139 --rc geninfo_all_blocks=1 00:42:27.139 --rc geninfo_unexecuted_blocks=1 00:42:27.139 00:42:27.139 ' 00:42:27.139 09:18:17 -- common/autotest_common.sh@1702 -- $ LCOV_OPTS=' 00:42:27.139 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:27.139 --rc genhtml_branch_coverage=1 00:42:27.139 --rc genhtml_function_coverage=1 00:42:27.139 --rc genhtml_legend=1 00:42:27.139 --rc geninfo_all_blocks=1 00:42:27.139 --rc geninfo_unexecuted_blocks=1 00:42:27.139 00:42:27.139 ' 00:42:27.139 09:18:17 -- common/autotest_common.sh@1703 -- $ export 'LCOV=lcov 00:42:27.139 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:27.139 --rc genhtml_branch_coverage=1 00:42:27.139 --rc genhtml_function_coverage=1 00:42:27.139 --rc genhtml_legend=1 00:42:27.139 --rc geninfo_all_blocks=1 00:42:27.139 --rc geninfo_unexecuted_blocks=1 00:42:27.139 00:42:27.139 ' 00:42:27.139 09:18:17 -- common/autotest_common.sh@1703 -- $ LCOV='lcov 00:42:27.139 --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 00:42:27.139 --rc genhtml_branch_coverage=1 00:42:27.139 --rc genhtml_function_coverage=1 00:42:27.139 --rc genhtml_legend=1 00:42:27.139 --rc geninfo_all_blocks=1 00:42:27.139 --rc geninfo_unexecuted_blocks=1 00:42:27.139 00:42:27.139 ' 00:42:27.139 09:18:17 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:42:27.139 09:18:17 -- scripts/common.sh@15 -- $ shopt -s extglob 00:42:27.139 09:18:17 -- scripts/common.sh@544 -- $ [[ -e /bin/wpdk_common.sh ]] 00:42:27.139 09:18:17 -- scripts/common.sh@552 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:42:27.139 09:18:17 -- scripts/common.sh@553 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:42:27.139 09:18:17 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:27.139 09:18:17 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:27.139 09:18:17 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:27.139 09:18:17 -- paths/export.sh@5 -- $ export PATH 00:42:27.139 09:18:17 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:42:27.139 09:18:17 -- common/autobuild_common.sh@485 -- $ out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:42:27.139 09:18:17 -- common/autobuild_common.sh@486 -- $ date +%s 00:42:27.140 09:18:17 -- common/autobuild_common.sh@486 -- $ mktemp -dt spdk_1730881097.XXXXXX 00:42:27.140 09:18:17 -- common/autobuild_common.sh@486 -- $ SPDK_WORKSPACE=/tmp/spdk_1730881097.NTajpN 00:42:27.140 09:18:17 -- common/autobuild_common.sh@488 -- $ [[ -n '' ]] 00:42:27.140 09:18:17 -- common/autobuild_common.sh@492 -- $ '[' -n '' ']' 00:42:27.140 09:18:17 -- common/autobuild_common.sh@495 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/' 00:42:27.140 09:18:17 -- common/autobuild_common.sh@499 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp' 00:42:27.140 09:18:17 -- common/autobuild_common.sh@501 -- $ scanbuild='scan-build -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:42:27.140 09:18:17 -- common/autobuild_common.sh@502 -- $ get_config_params 00:42:27.140 09:18:17 -- common/autotest_common.sh@407 -- $ xtrace_disable 00:42:27.140 09:18:17 -- common/autotest_common.sh@10 -- $ set +x 00:42:27.140 09:18:17 -- common/autobuild_common.sh@502 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user' 00:42:27.140 09:18:17 -- common/autobuild_common.sh@504 -- $ start_monitor_resources 00:42:27.140 09:18:17 -- pm/common@17 -- $ local monitor 00:42:27.140 09:18:17 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:42:27.140 09:18:17 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:42:27.140 09:18:17 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:42:27.140 09:18:17 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:42:27.140 09:18:17 -- pm/common@21 -- $ date +%s 00:42:27.140 09:18:17 -- pm/common@25 -- $ sleep 1 00:42:27.140 09:18:17 -- pm/common@21 -- $ date +%s 00:42:27.140 09:18:17 -- pm/common@21 -- $ date +%s 00:42:27.140 09:18:17 -- pm/common@21 -- $ date +%s 00:42:27.140 09:18:17 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1730881097 00:42:27.140 09:18:17 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1730881097 00:42:27.140 09:18:17 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1730881097 00:42:27.140 09:18:17 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1730881097 00:42:27.140 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1730881097_collect-cpu-load.pm.log 00:42:27.140 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1730881097_collect-vmstat.pm.log 00:42:27.401 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1730881097_collect-cpu-temp.pm.log 00:42:27.401 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1730881097_collect-bmc-pm.bmc.pm.log 00:42:28.343 09:18:18 -- common/autobuild_common.sh@505 -- $ trap stop_monitor_resources EXIT 00:42:28.343 09:18:18 -- spdk/autopackage.sh@10 -- $ [[ 0 -eq 1 ]] 00:42:28.343 09:18:18 -- spdk/autopackage.sh@14 -- $ timing_finish 00:42:28.343 09:18:18 -- common/autotest_common.sh@736 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:42:28.343 09:18:18 -- common/autotest_common.sh@737 -- $ [[ -x /usr/local/FlameGraph/flamegraph.pl ]] 00:42:28.343 09:18:18 -- common/autotest_common.sh@740 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:42:28.343 09:18:18 -- spdk/autopackage.sh@1 -- $ stop_monitor_resources 00:42:28.343 09:18:18 -- pm/common@29 -- $ signal_monitor_resources TERM 00:42:28.343 09:18:18 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:42:28.343 09:18:18 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:42:28.343 09:18:18 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:42:28.343 09:18:18 -- pm/common@44 -- $ pid=2238079 00:42:28.343 09:18:18 -- pm/common@50 -- $ kill -TERM 2238079 00:42:28.343 09:18:18 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:42:28.343 09:18:18 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:42:28.343 09:18:18 -- pm/common@44 -- $ pid=2238080 00:42:28.343 09:18:18 -- pm/common@50 -- $ kill -TERM 2238080 00:42:28.343 09:18:18 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:42:28.343 09:18:18 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:42:28.343 09:18:18 -- pm/common@44 -- $ pid=2238082 00:42:28.343 09:18:18 -- pm/common@50 -- $ kill -TERM 2238082 00:42:28.343 09:18:18 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:42:28.343 09:18:18 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:42:28.343 09:18:18 -- pm/common@44 -- $ pid=2238105 00:42:28.343 09:18:18 -- pm/common@50 -- $ sudo -E kill -TERM 2238105 00:42:28.343 + [[ -n 1566672 ]] 00:42:28.343 + sudo kill 1566672 00:42:28.354 [Pipeline] } 00:42:28.370 [Pipeline] // stage 00:42:28.376 [Pipeline] } 00:42:28.390 [Pipeline] // timeout 00:42:28.396 [Pipeline] } 00:42:28.410 [Pipeline] // catchError 00:42:28.415 [Pipeline] } 00:42:28.430 [Pipeline] // wrap 00:42:28.436 [Pipeline] } 00:42:28.450 [Pipeline] // catchError 00:42:28.459 [Pipeline] stage 00:42:28.461 [Pipeline] { (Epilogue) 00:42:28.474 [Pipeline] catchError 00:42:28.476 [Pipeline] { 00:42:28.489 [Pipeline] echo 00:42:28.491 Cleanup processes 00:42:28.498 [Pipeline] sh 00:42:28.788 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:42:28.788 2238231 /usr/bin/ipmitool sdr dump /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/sdr.cache 00:42:28.788 2238778 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:42:28.804 [Pipeline] sh 00:42:29.093 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:42:29.093 ++ grep -v 'sudo pgrep' 00:42:29.093 ++ awk '{print $1}' 00:42:29.093 + sudo kill -9 2238231 00:42:29.106 [Pipeline] sh 00:42:29.395 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:42:41.659 [Pipeline] sh 00:42:41.948 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:42:41.948 Artifacts sizes are good 00:42:41.963 [Pipeline] archiveArtifacts 00:42:41.983 Archiving artifacts 00:42:42.143 [Pipeline] sh 00:42:42.491 + sudo chown -R sys_sgci: /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:42:42.507 [Pipeline] cleanWs 00:42:42.518 [WS-CLEANUP] Deleting project workspace... 00:42:42.518 [WS-CLEANUP] Deferred wipeout is used... 00:42:42.526 [WS-CLEANUP] done 00:42:42.528 [Pipeline] } 00:42:42.545 [Pipeline] // catchError 00:42:42.556 [Pipeline] sh 00:42:42.843 + logger -p user.info -t JENKINS-CI 00:42:42.853 [Pipeline] } 00:42:42.867 [Pipeline] // stage 00:42:42.872 [Pipeline] } 00:42:42.886 [Pipeline] // node 00:42:42.891 [Pipeline] End of Pipeline 00:42:42.935 Finished: SUCCESS